To Monitor or to Trust: Observing Robot's Behavior based on a Game-Theoretic Model of Trust

In scenarios where a robot generates and executes a plan, there may be instances where this generated plan is less costly for the robot to execute but incomprehensible to the human. When the human acts as a supervisor and is held accountable for the robot's plan, the human may be at a higher risk if the incomprehensible behavior is deemed to be infeasible or unsafe. In such cases, the robot, who may be unaware of the human's exact expectations, may choose to execute (1) the most constrained plan (i.e. one preferred by all possible supervisors) incurring the added cost of executing highly sub-optimal behavior when the human is monitoring it and (2) deviate to a more optimal plan when the human looks away. While robots do not have human-like ulterior motives (such as being lazy), such behavior may occur because the robot has to cater to the needs of different human supervisors. In such settings, the robot, being a rational agent, should take any chance it gets to deviate to a lower cost plan. On the other hand, continuous monitoring of the robot's behavior is often difficult for humans because it costs them valuable resources (e.g., time, cognitive overload, etc.). Thus, to optimize the cost for monitoring while ensuring the robots follow the safe behavior, we model this problem in the game-theoretic framework of trust. In settings where the human does not initially trust the robot, pure-strategy Nash Equilibrium provides a useful policy for the human.

[1]  Siddhartha S. Srinivasa,et al.  Legibility and predictability of robot motion , 2013, 2013 8th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[2]  Tom Bylander,et al.  The Computational Complexity of Propositional STRIPS Planning , 1994, Artif. Intell..

[3]  Milind Tambe,et al.  From physical security to cybersecurity , 2015, J. Cybersecur..

[4]  Sarit Kraus,et al.  Playing games for security: an efficient exact algorithm for solving Bayesian Stackelberg games , 2008, AAMAS.

[5]  Sailik Sengupta,et al.  A Game Theoretic Approach to Strategy Generation for Moving Target Defense in Web Applications , 2017, AAMAS.

[6]  Subbarao Kambhampati,et al.  Robust planning with incomplete domain models , 2017, Artif. Intell..

[7]  Shambhu J. Upadhyaya,et al.  Towards Modeling Trust Based Decisions: A Game Theoretic Approach , 2007, ESORICS.

[8]  Lars Karlsson,et al.  Grandpa Hates Robots - Interaction Constraints for Planning in Inhabited Environments , 2014, AAAI.

[9]  Yu Zhang,et al.  Explicable Robot Planning as Minimizing Distance from Expected Behavior , 2016, ArXiv.

[10]  Sailik Sengupta Moving Target Defense: A Symbiotic Framework for AI & Security , 2017, AAMAS.

[11]  Malte Helmert,et al.  The Fast Downward Planning System , 2006, J. Artif. Intell. Res..

[12]  Subbarao Kambhampati,et al.  Explanations as Model Reconciliation - A Multi-Agent Perspective , 2017, AAAI Fall Symposia.

[13]  Haifeng Xu,et al.  Deceiving Cyber Adversaries: A Game Theoretic Approach , 2018, AAMAS.

[14]  Anca D. Dragan,et al.  The Off-Switch Game , 2016, IJCAI.

[15]  Yu Zhang,et al.  Plan explicability and predictability for robot task planning , 2015, 2017 IEEE International Conference on Robotics and Automation (ICRA).