暂无分享,去创建一个
Stefano Teso | Mohit Kumar | Teodora Popordanoska | Stefano Teso | Mohit Kumar | Teodora Popordanoska
[1] Jude W. Shavlik,et al. in Advances in Neural Information Processing , 1996 .
[2] Panagiotis Papapetrou,et al. A peek into the black box: exploring classifiers by randomization , 2014, Data Mining and Knowledge Discovery.
[3] Xiaojin Zhu,et al. Machine Teaching: An Inverse Problem to Machine Learning and an Approach Toward Optimal Education , 2015, AAAI.
[4] Albert Gordo,et al. Learning Global Additive Explanations for Neural Nets Using Model Distillation , 2018 .
[5] Thomas G. Dietterich. Steps Toward Robust Artificial Intelligence , 2017, AI Mag..
[6] David Maxwell Chickering,et al. ICE: Enabling Non-Experts to Build Models Interactively for Large-Scale Lopsided Problems , 2014, ArXiv.
[7] Kristian Kersting,et al. Right for the Wrong Scientific Reasons: Revising Deep Networks by Interacting with their Explanations , 2020, ArXiv.
[8] Samuel J. Gershman,et al. Human-in-the-Loop Interpretability Prior , 2018, NeurIPS.
[9] Kristian Kersting,et al. Explanatory Interactive Machine Learning , 2019, AIES.
[10] Panagiotis G. Ipeirotis,et al. Beat the Machine: Challenging Humans to Find a Predictive Model's “Unknown Unknowns” , 2015, JDIQ.
[11] João Gama,et al. A survey on concept drift adaptation , 2014, ACM Comput. Surv..
[12] Bin Li,et al. A survey on instance selection for active learning , 2012, Knowledge and Information Systems.
[13] Albert Gordo,et al. Transparent Model Distillation , 2018, ArXiv.
[14] Lars Niklasson,et al. Accuracy vs. comprehensibility in data mining models , 2004 .
[15] Edith Law,et al. Contradict the Machine: A Hybrid Approach to Identifying Unknown Unknowns , 2019, AAMAS.
[16] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.
[17] Maya Cakmak,et al. Eliciting good teaching from humans for machine learners , 2014, Artif. Intell..
[18] Pietro Perona,et al. Near-Optimal Machine Teaching via Explanatory Teaching Sets , 2018, AISTATS.
[19] Stefano Teso. Toward Faithful Explanatory Active Learning with Self-explainable Neural Nets , 2019 .
[20] Osbert Bastani,et al. Interpreting Blackbox Models via Model Extraction , 2017, ArXiv.
[21] Sanjay Ranka,et al. Global Model Interpretation Via Recursive Partitioning , 2018, 2018 IEEE 20th International Conference on High Performance Computing and Communications; IEEE 16th International Conference on Smart City; IEEE 4th International Conference on Data Science and Systems (HPCC/SmartCity/DSS).
[22] Cynthia Rudin,et al. Interpretable Image Recognition with Hierarchical Prototypes , 2019, HCOMP.
[23] Sanjoy Dasgupta,et al. Teaching a black-box learner , 2019, ICML.
[24] Steve Hanneke,et al. Theory of Disagreement-Based Active Learning , 2014, Found. Trends Mach. Learn..
[25] Eric Horvitz,et al. Identifying Unknown Unknowns in the Open World: Representations and Policies for Guided Exploration , 2016, AAAI.
[26] Fausto Giunchiglia,et al. Fixing Mislabeling by Human Annotators Leveraging Conflict Resolution and Prior Knowledge , 2019, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..
[27] T. Kathirvalavakumar,et al. Reverse Engineering the Neural Networks for Rule Extraction in Classification Problems , 2011, Neural Processing Letters.
[28] Manuel Lopes,et al. Interactive Optimal Teaching with Unknown Learners , 2018, IJCAI.
[29] Terrance E. Boult,et al. Learning and the Unknown: Surveying Steps toward Open World Recognition , 2019, AAAI.
[30] Giles Hooker,et al. Tree Space Prototypes: Another Look at Making Tree Ensembles Interpretable , 2016, FODS.
[31] Zhiyuan Liu,et al. Incentivized Exploration for Multi-Armed Bandits under Reward Drift , 2020, AAAI.
[32] David Cohn,et al. Active Learning , 2010, Encyclopedia of Machine Learning.
[33] Franco Turini,et al. A Survey of Methods for Explaining Black Box Models , 2018, ACM Comput. Surv..
[34] John Langford,et al. Search Improves Label for Active Learning , 2016, NIPS.
[35] Mark Craven,et al. Understanding Learned Models by Identifying Important Features at the Right Resolution , 2018, AAAI.
[36] Rich Caruana,et al. Model compression , 2006, KDD '06.
[37] Eric Horvitz,et al. Beyond Accuracy: The Role of Mental Models in Human-AI Team Performance , 2019, HCOMP.
[38] Klaus-Robert Müller,et al. Explanations can be manipulated and geometry is to blame , 2019, NeurIPS.
[39] Olcay Boz,et al. Extracting decision trees from trained neural networks , 2002, KDD.
[40] Andrew P. Bradley,et al. Rule extraction from support vector machines: A review , 2010, Neurocomputing.
[41] Joachim Diederich,et al. Survey and critique of techniques for extracting rules from trained artificial neural networks , 1995, Knowl. Based Syst..
[42] Joachim Diederich,et al. Rule Extraction from Support Vector Machines , 2008, Studies in Computational Intelligence.
[43] Foster J. Provost,et al. Why label when you can search?: alternatives to active learning for applying human resources to build classification models under extreme class imbalance , 2010, KDD.