STARdom: an architecture for trusted and secure human-centered manufacturing systems

[1]  Jože M. Rožanec,et al.  Explainable Demand Forecasting: A Data Mining Goldmine , 2021, WWW.

[2]  Paul Dourish,et al.  Directive Explanations for Actionable Explainability in Machine Learning Applications , 2021, ACM Trans. Interact. Intell. Syst..

[3]  Gyogwon Koo,et al.  Automated defect inspection system for metal surfaces based on deep learning and data augmentation , 2020 .

[4]  Asaf Shabtai,et al.  When Explainability Meets Adversarial Learning: Detecting Adversarial Examples using SHAP Signatures , 2019, 2020 International Joint Conference on Neural Networks (IJCNN).

[5]  S. Nahavandi Industry 5.0—A Human-Centric Solution , 2019, Sustainability.

[6]  Ross Maciejewski,et al.  Explaining Vulnerabilities to Adversarial Machine Learning through Visual Analytics , 2019, IEEE Transactions on Visualization and Computer Graphics.

[7]  John Soldatos,et al.  The Digital Shopfloor: Industrial Automation in the Industry 4.0 Era , 2019, The Digital Shopfloor: Industrial Automation in the Industry 4.0 Era.

[8]  K. Müller,et al.  Unmasking Clever Hans predictors and assessing what machines really learn , 2019, Nature Communications.

[9]  Jiliang Zhang,et al.  Adversarial Examples: Opportunities and Challenges , 2018, IEEE Transactions on Neural Networks and Learning Systems.

[10]  Reiner Anderl,et al.  Topological Approach for Mapping Technologies in Reference Architectural Model Industrie 4.0 (RAMI 4.0) , 2017 .

[11]  Alberto Tellaeche,et al.  Natural multimodal communication for human–robot collaboration , 2017 .

[12]  Scott Lundberg,et al.  A Unified Approach to Interpreting Model Predictions , 2017, NIPS.

[13]  John Schulman,et al.  Concrete Problems in AI Safety , 2016, ArXiv.

[14]  Francesco Ricci,et al.  A survey of active learning in collaborative filtering recommender systems , 2016, Comput. Sci. Rev..

[15]  Carlos Guestrin,et al.  "Why Should I Trust You?": Explaining the Predictions of Any Classifier , 2016, ArXiv.

[16]  Nan Hu,et al.  SHERLOCK: Simple Human Experiments Regarding Locally Observed Collective Knowledge , 2015 .

[17]  D. L. Métayer,et al.  A Multi-layered Approach for Tailored Black-Box Explanations , 2020, ICPR Workshops.

[18]  Francisco S. Melo,et al.  Learning from Explanations and Demonstrations: A Pilot Study , 2020, NL4XAI.

[19]  High-Level Expert Group on Artificial Intelligence – Draft Ethics Guidelines for Trustworthy AI , 2019 .

[20]  Burr Settles,et al.  Active Learning Literature Survey , 2009 .