Formal model of human erroneous behavior for safety analysis in collaborative robotics

Abstract Recent developments in manufacturing technologies, also known as Industry 4.0, seek to build Smart Factories where supply chains and production lines are equipped with a higher level of automation. However, this significant innovation does not entirely eliminate the need for the presence of human operators; on the contrary, it requires them to collaborate with robots and execute hybrid tasks. Thus, creating safe workspaces for human operators is crucial for the future of factories where humans and robots collaborate closely in common workspaces. The uncertainty of human behavior and, consequently, of the actual execution of workflows, pose significant challenges to the safety of collaborative applications. This paper extends our earlier work, a formal verification methodology to analyze the safety of collaborative robotics applications (Askarpour et al. 2017) [1], with a rich non-deterministic formal model of operator behaviors that captures the hazardous situations resulting from human errors. The model allows safety engineers to refine their designs until all plausible erroneous behaviors are considered and mitigated. The solidity of the proposed approach is evaluated on a pair of real-life case studies.

[1]  Matthew L. Bolton,et al.  Model Checking Human-Human Communication Protocols Using Task Models and Miscommunication Generation , 2015, J. Aerosp. Inf. Syst..

[2]  Matthew L. Bolton,et al.  Properties for formally assessing the performance level of human-human collaborative procedures with miscommunications and erroneous human behavior ☆ , 2016 .

[3]  Philippe A. Palanque,et al.  Task Model-Based Systematic Analysis of Both System Failures and Human Errors , 2016, IEEE Transactions on Human-Machine Systems.

[4]  Ellen J. Bass,et al.  Using Formal Verification to Evaluate Human-Automation Interaction: A Review , 2013, IEEE Transactions on Systems, Man, and Cybernetics: Systems.

[5]  Sotiris Makris,et al.  Design Considerations for Safe Human-robot Collaborative Workplaces , 2015 .

[6]  Richard A. Wysk,et al.  Formal model of human material-handling tasks for control of manufacturing systems , 2006, IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans.

[7]  Luciano Baresi,et al.  Efficient Scalable Verification of LTL Specifications , 2015, 2015 IEEE/ACM 37th IEEE International Conference on Software Engineering.

[8]  Fabio Paternò,et al.  Preventing user errors by systematic analysis of deviations from the system task model , 2002, Int. J. Hum. Comput. Stud..

[9]  Richard Greenough,et al.  A decision support tool based on QFD and FMEA for the selection of manufacturing automation technologies , 2008 .

[10]  Dino Mandrioli,et al.  Modeling time in computing: A taxonomy and a comparative survey , 2008, CSUR.

[11]  Michael D. Harrison,et al.  THEA: A Technique for Human Error Assessment Early in Design , 2001, INTERACT.

[12]  Dino Mandrioli,et al.  Modeling Operator Behavior in the Safety Analysis of Collaborative Robotic Applications , 2017, SAFECOMP.

[13]  A. D. Swain,et al.  Handbook of human-reliability analysis with emphasis on nuclear power plant applications. Final report , 1983 .

[14]  Marco Pistore,et al.  NuSMV 2: An OpenSource Tool for Symbolic Model Checking , 2002, CAV.

[15]  Dino Mandrioli,et al.  A Human-in-the-Loop Perspective for Safety Assessment in Robotic Applications , 2017, Ershov Informatics Conference.

[16]  Frank J. Lee,et al.  Simple cognitive modeling in a complex cognitive architecture , 2003, CHI '03.

[17]  J. R. Fragola,et al.  Human reliability analysis , 1988 .

[18]  Titanilla Komenda,et al.  A Morphology of Human Robot Collaboration Systems for Industrial Assembly , 2018 .

[19]  J.C. Williams,et al.  A data-based method for assessing and reducing human error to improve operational performance , 1988, Conference Record for 1988 IEEE Fourth Conference on Human Factors and Power Plants,.

[20]  Peter A. Lindsay,et al.  Modelling Erroneous Operator Behaviours for an Air-Traffic Control Task , 2002, AUIC.

[21]  Lorrie Faith Cranor,et al.  A Framework for Reasoning About the Human in the Loop , 2008, UPSEC.

[22]  Matteo Pradella,et al.  Bounded satisfiability checking of metric temporal logic specifications , 2013, TSEM.

[23]  Terrence Fong,et al.  A Survey of Methods for Safe Human-Robot Interaction , 2017, Found. Trends Robotics.

[24]  John E. Laird,et al.  The Soar Cognitive Architecture , 2012 .

[25]  Namhun Kim,et al.  An affordance-based formalism for modeling human-involvement in complex systems for prospective control , 2010, Proceedings of the 2010 Winter Simulation Conference.

[26]  Erik Hollnagel,et al.  Cognitive reliability and error analysis method : CREAM , 1998 .

[27]  Alan J. Dix,et al.  Physigrams: modelling devices for natural interaction , 2009, Formal Aspects of Computing.

[28]  Dino Mandrioli,et al.  SAFER-HRC: Safety Analysis Through Formal vERification in Human-Robot Collaboration , 2016, SAFECOMP.

[29]  Peter A. Lindsay,et al.  Formal analysis of human-computer interaction using model-checking , 2005, Third IEEE International Conference on Software Engineering and Formal Methods (SEFM'05).

[30]  Ann Blandford,et al.  From a Formal User Model to Design Rules , 2002, DSV-IS.

[31]  Ellen J. Bass,et al.  A Systematic Approach to Model Checking Human–Automation Interaction Using Task Analytic Models , 2011, IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans.

[32]  Fabio Paternò,et al.  ConcurTaskTrees: A Diagrammatic Notation for Specifying Task Models , 1997, INTERACT.

[33]  George Chryssolouris,et al.  An efficient approach to human motion modeling for the verification of human-centric product design and manufacturing in virtual environments , 2007 .

[34]  Ellen J. Bass,et al.  Formally verifying human–automation interaction as part of a system model: limitations and tradeoffs , 2010, Innovations in Systems and Software Engineering.

[35]  Antonio C. Siochi,et al.  The UAN: a user-oriented representation for direct manipulation interface designs , 1990, TOIS.

[36]  B Kirwan,et al.  Human error identification techniques for risk assessment of high risk systems--Part 1: Review and evaluation of techniques. , 1998, Applied ergonomics.

[37]  Jeremy A Marvel,et al.  Implementing Speed and Separation Monitoring in Collaborative Robot Workcells. , 2017, Robotics and computer-integrated manufacturing.

[38]  Ann Blandford,et al.  Formally Justifying User-Centred Design Rules: A Case Study on Post-completion Errors , 2004, IFM.

[39]  Ellen J. Bass,et al.  Generating phenotypical erroneous human behavior to evaluate human-automation interaction using model checking , 2012, Int. J. Hum. Comput. Stud..