暂无分享,去创建一个
Geoffrey E. Hinton | Rich Caruana | Rishabh Agarwal | Xuezhou Zhang | Nicholas Frosst | Rishabh Agarwal | R. Caruana | Nicholas Frosst | Xuezhou Zhang
[1] E. Mazzoni,et al. An interpretable bimodal neural network characterizes the sequence and preexisting chromatin predictors of induced transcription factor binding , 2021, Genome Biology.
[2] Yoshua Bengio,et al. On the Spectral Bias of Neural Networks , 2018, ICML.
[3] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[4] R. Tibshirani,et al. Generalized Additive Models , 1986 .
[5] Gianluca Bontempi,et al. Adaptive Machine Learning for Credit Card Fraud Detection , 2015 .
[6] Geoffrey E. Hinton,et al. Learning representations by back-propagating errors , 1986, Nature.
[7] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[8] Yoshua Bengio,et al. A Closer Look at Memorization in Deep Networks , 2017, ICML.
[9] D. Sculley,et al. Google Vizier: A Service for Black-Box Optimization , 2017, KDD.
[10] Trevor Hastie,et al. Generalized linear and generalized additive models in studies of species distributions: setting the scene , 2002 .
[11] R. Souza,et al. A case study of hurdle and generalized additive models in astronomy: the escape of ionizing radiation , 2018, Monthly Notices of the Royal Astronomical Society.
[12] Ilya Sutskever,et al. Language Models are Unsupervised Multitask Learners , 2019 .
[13] Lev V. Utkin,et al. SurvNAM: The machine learning survival model explanation , 2021, Neural Networks.
[14] The AI revolution in scientific research , 2019 .
[15] Christopher T. Lowenkamp,et al. False Positives, False Negatives, and False Analyses: A Rejoinder to "Machine Bias: There's Software Used across the Country to Predict Future Criminals. and It's Biased against Blacks" , 2016 .
[16] T. H. Kyaw,et al. Multiparameter Intelligent Monitoring in Intensive Care II: A public-access intensive care unit database* , 2011, Critical care medicine.
[17] Johannes Gehrke,et al. Intelligible Models for HealthCare: Predicting Pneumonia Risk and Hospital 30-day Readmission , 2015, KDD.
[18] R. Tibshirani,et al. Generalized additive models for medical research , 1995, Statistical methods in medical research.
[19] Tim Appenzeller,et al. The AI revolution in science , 2017 .
[20] Ribana Roscher,et al. Explainable Machine Learning for Scientific Insights and Discoveries , 2019, IEEE Access.
[21] Rich Caruana,et al. Multitask Learning , 1998, Encyclopedia of Machine Learning and Data Mining.
[22] M. Saeed,et al. Multiparameter Intelligent Monitoring in Intensive Care Ii (Mimic-Ii): A Public-Access Intensive Care Unit Database , 2011 .
[23] Tianqi Chen,et al. XGBoost: A Scalable Tree Boosting System , 2016, KDD.
[24] Cynthia Rudin,et al. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead , 2018, Nature Machine Intelligence.
[25] Jian Sun,et al. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[26] Johannes Gehrke,et al. Accurate intelligible models with pairwise interactions , 2013, KDD.
[27] William J. E. Potts,et al. Generalized additive neural networks , 1999, KDD '99.
[28] Gaël Varoquaux,et al. Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..
[29] Rich Caruana,et al. Distill-and-Compare: Auditing Black-Box Models Using Transparent Model Distillation , 2017, AIES.
[30] R. Pace,et al. Sparse spatial autoregressions , 1997 .
[31] Rich Caruana,et al. InterpretML: A Unified Framework for Machine Learning Interpretability , 2019, ArXiv.
[32] Geoffrey E. Hinton,et al. Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.
[33] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[34] M. Kearns,et al. Fairness in Criminal Justice Risk Assessments: The State of the Art , 2017, Sociological Methods & Research.
[35] Hany Farid,et al. The accuracy, fairness, and limits of predicting recidivism , 2018, Science Advances.
[36] Kurt Hornik,et al. Multilayer feedforward networks are universal approximators , 1989, Neural Networks.
[37] Johannes Gehrke,et al. Intelligible models for classification and regression , 2012, KDD.
[38] Jonathan T. Barron,et al. Fourier Features Let Networks Learn High Frequency Functions in Low Dimensional Domains , 2020, NeurIPS.
[39] Jasper Snoek,et al. Practical Bayesian Optimization of Machine Learning Algorithms , 2012, NIPS.
[40] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[41] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[42] P. Baldi,et al. Development and validation of an interpretable neural network for prediction of postoperative in-hospital mortality , 2021, npj Digital Medicine.
[43] A. Krizhevsky. Convolutional Deep Belief Networks on CIFAR-10 , 2010 .
[44] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.
[45] Cynthia Rudin,et al. Interpretable Machine Learning: Fundamental Principles and 10 Grand Challenges , 2021, ArXiv.