Feature Importance Measurement based on Decision Tree Sampling
暂无分享,去创建一个
[1] C. Rudin,et al. FasterRisk: Fast and Accurate Interpretable Risk Scores , 2022, NeurIPS.
[2] I. Takeuchi,et al. A Confidence Machine for Sparse High-Order Interaction Model , 2022, ArXiv.
[3] Ichiro Takeuchi,et al. Fast and More Powerful Selective Inference for Sparse High-order Interaction Model , 2021, AAAI.
[4] Cesare Tinelli,et al. Handbook of Satisfiability , 2021, Handbook of Satisfiability.
[5] Wei-Yin Loh,et al. Classification and regression trees , 2011, WIREs Data Mining Knowl. Discov..
[6] Siegfried Nijssen,et al. Learning optimal decision trees using constraint programming , 2020, Constraints.
[7] Mikoláš Janota,et al. SAT-Based Encodings for Optimal Decision Trees with Explicit Paths , 2020, SAT.
[8] Mate Soos,et al. Tinted, Detached, and Lazy CNF-XOR Solving and Its Applications to Counting and Sampling , 2020, CAV.
[9] C. Rudin,et al. In Pursuit of Interpretable, Fair and Accurate Machine Learning for Criminal Recidivism Prediction , 2020, Journal of Quantitative Criminology.
[10] Amina Adadi,et al. Explainable AI for Healthcare: From Black Box to Interpretable Models , 2020 .
[11] Koji Tsuda,et al. An interpretable machine learning model for diagnosis of Alzheimer's disease , 2019, PeerJ.
[12] Cynthia Rudin,et al. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead , 2018, Nature Machine Intelligence.
[13] Joao Marques-Silva,et al. Learning Optimal Decision Trees with SAT , 2018, IJCAI.
[14] Ankur Teredesai,et al. Interpretable Machine Learning in Healthcare , 2018, 2018 IEEE International Conference on Healthcare Informatics (ICHI).
[15] Koushik Sen,et al. Efficient Sampling of SAT Solutions for Testing , 2018, 2018 IEEE/ACM 40th International Conference on Software Engineering (ICSE).
[16] Shamim Nemati,et al. An Interpretable Machine Learning Model for Accurate Prediction of Sepsis in the ICU , 2017, Critical care medicine.
[17] Margo I. Seltzer,et al. Learning Certifiably Optimal Rule Lists , 2017, KDD.
[18] Jure Leskovec,et al. Interpretable Decision Sets: A Joint Framework for Description and Prediction , 2016, KDD.
[19] Seth Flaxman,et al. European Union Regulations on Algorithmic Decision-Making and a "Right to Explanation" , 2016, AI Mag..
[20] Cynthia Rudin,et al. Interpretable classifiers using rules and Bayesian analysis: Building a better stroke prediction model , 2015, ArXiv.
[21] Barry O'Sullivan,et al. Minimising Decision Tree Size as Combinatorial Optimisation , 2009, CP.
[22] Nikolaj Bjørner,et al. Z3: An Efficient SMT Solver , 2008, TACAS.
[23] Jianqing Fan,et al. Sure independence screening for ultrahigh dimensional feature space , 2006, math/0612857.
[24] Edward I. George,et al. Bayesian Treed Models , 2002, Machine Learning.
[25] L. Breiman. Random Forests , 2001, Encyclopedia of Machine Learning and Data Mining.
[26] H. Chipman,et al. Bayesian CART Model Search , 1998 .
[27] Adrian F. M. Smith,et al. A Bayesian CART algorithm , 1998 .
[28] Y. C. Pati,et al. Orthogonal matching pursuit: recursive function approximation with applications to wavelet decomposition , 1993, Proceedings of 27th Asilomar Conference on Signals, Systems and Computers.
[29] J. Ross Quinlan,et al. C4.5: Programs for Machine Learning , 1992 .
[30] J. Ross Quinlan,et al. Induction of Decision Trees , 1986, Machine Learning.
[31] K. Tsuda,et al. Self-learning Entropic Population Annealing for Interpretable Materials Design , 2022, Digital Discovery.
[32] H. Chipman,et al. Bayesian Additive Regression Trees , 2006 .
[33] R. Tibshirani. Regression Shrinkage and Selection via the Lasso , 1996 .
[34] G. S. Tseitin. On the Complexity of Derivation in Propositional Calculus , 1983 .