Anomaly Attribution with Likelihood Compensation
暂无分享,去创建一个
Amit Dhurandhar | Tsuyoshi Idé | Naoki Abe | Jirí Navrátil | Moninder Singh | N. Abe | Moninder Singh | Amit Dhurandhar | Jirí Navrátil | T. Idé
[1] Manish Marwah,et al. ACE – An Anomaly Contribution Explainer for Cyber-Security Applications , 2019, 2019 IEEE International Conference on Big Data (Big Data).
[2] Thomas G. Dietterich,et al. Anomaly detection in the presence of missing values for weather data quality control , 2018, COMPASS.
[3] D. Ruppert. The Elements of Statistical Learning: Data Mining, Inference, and Prediction , 2004 .
[4] Carlos Guestrin,et al. Anchors: High-Precision Model-Agnostic Explanations , 2018, AAAI.
[5] Chris Russell,et al. Counterfactual Explanations Without Opening the Black Box: Automated Decisions and the GDPR , 2017, ArXiv.
[6] Ioana Giurgiu,et al. Additive Explanations for Anomalies Detected from Multivariate Temporal Data , 2019, CIKM.
[7] Bernd Bischl,et al. Visualizing the Feature Importance for Black Box Models , 2018, ECML/PKDD.
[8] Brandon M. Greenwell,et al. Interpretable Machine Learning , 2019, Hands-On Machine Learning with R.
[9] M. de Rijke,et al. Why does my model fail?: contrastive local explanations for retail forecasting , 2019, FAT*.
[10] Fei Tao,et al. Digital twin-driven product design, manufacturing and service with big data , 2017, The International Journal of Advanced Manufacturing Technology.
[11] Erik Strumbelj,et al. An Efficient Explanation of Individual Classifications using Game Theory , 2010, J. Mach. Learn. Res..
[12] John Sipple,et al. Interpretable, Multidimensional, Multimodal Anomaly Detection with Negative Sampling for Detection of Device Failure , 2020, ICML.
[13] Vivekananda Roy,et al. Selection of Tuning Parameters, Solution Paths and Standard Errors for Bayesian Lassos , 2017 .
[14] Milos Manic,et al. Toward Explainable Deep Neural Network Based Anomaly Detection , 2018, 2018 11th International Conference on Human System Interaction (HSI).
[15] Amit Dhurandhar,et al. Explanations based on the Missing: Towards Contrastive Explanations with Pertinent Negatives , 2018, NeurIPS.
[16] Stephen P. Boyd,et al. Proximal Algorithms , 2013, Found. Trends Optim..
[17] Erik Strumbelj,et al. Explaining prediction models and individual predictions with feature contributions , 2014, Knowledge and Information Systems.
[18] Andrea Vedaldi,et al. Interpretable Explanations of Black Boxes by Meaningful Perturbation , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[19] Ankur Taly,et al. Axiomatic Attribution for Deep Networks , 2017, ICML.
[20] Wojciech Samek,et al. Explainable AI: Interpreting, Explaining and Visualizing Deep Learning , 2019, Explainable AI.
[21] VARUN CHANDOLA,et al. Anomaly detection: A survey , 2009, CSUR.
[22] W. W. Muir,et al. Regression Diagnostics: Identifying Influential Data and Sources of Collinearity , 1980 .
[23] Alfredo Cuzzocrea,et al. Interpretable Anomaly Prediction: Predicting anomalous behavior in industry 4.0 settings via regularized logistic regression tools , 2020, Data Knowl. Eng..
[24] Ankur Taly,et al. Explainable AI in Industry: Practical Challenges and Lessons Learned , 2020, WWW.
[25] Graham J. Williams,et al. On-Line Unsupervised Outlier Detection Using Finite Mixtures with Discounting Learning Algorithms , 2000, KDD '00.
[26] Alexander Binder,et al. On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation , 2015, PloS one.
[27] Scott Lundberg,et al. A Unified Approach to Interpreting Model Predictions , 2017, NIPS.
[28] Carlos Guestrin,et al. "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.