Attributing Blame to Robots: I. The Influence of Robot Autonomy

Objective The research examined how humans attribute blame to humans, nonautonomous robots, autonomous robots, or environmental factors for scenarios in which errors occur. Background When robots and humans serve on teams, human perception of their technological team members can be a critical component of successful cooperation, especially when task completion fails. Methods Participants read a set of scenarios that described human–robot team task failures. Separate scenarios were written to emphasize the role of the human, the robot, or environmental factors in producing the task failure. After reading each scenario, the participants allocated blame for the failure among the human, robot, and environmental factors. Results In general, the order of amount of blame was humans, robots, and environmental factors. If the scenario described the robot as nonautonomous, the participants attributed almost as little blame to them as to the environmental factors; in contrast, if the scenario described the robot as autonomous, the participants attributed almost as much blame to them as to the human. Conclusion We suggest that humans use a hierarchy of blame in which robots are seen as partial social actors, with the degree to which people view them as social actors depending on the degree of autonomy. Application The acceptance of robots by human co-workers will be a function of the attribution of blame when errors occur in the workplace. The present research suggests that greater autonomy for the robot will result in greater attribution of blame in work tasks.

[1]  Mika Luimula,et al.  Social service robots in public and private environments , 2012 .

[2]  Christopher D. Wickens,et al.  A model for types and levels of human interaction with automation , 2000, IEEE Trans. Syst. Man Cybern. Part A.

[3]  Michael D. Buhrmester,et al.  Amazon's Mechanical Turk , 2011, Perspectives on psychological science : a journal of the Association for Psychological Science.

[4]  George A. Bekey,et al.  AUTONOMOUS ROBOTS, From Biological Inspiration to Implementation and Control, by G.A. Bekey, MIT Press, 2005, xv + 577 pp., index, ISBN 0-262-02578-7, 25 pages of references (Hb. £35.95) , 2005, Robotica.

[5]  Clifford Nass,et al.  Computers are social actors , 1994, CHI '94.

[6]  Evolutionary Social Psychology , 2010 .

[7]  Pamela J. Hinds,et al.  Who Should I Blame? Effects of Autonomy and Transparency on Attributions in Human-Robot Interaction , 2006, ROMAN 2006 - The 15th IEEE International Symposium on Robot and Human Interactive Communication.

[8]  E. Deci,et al.  Self-regulation and the problem of human autonomy: does psychology need choice, self-determination, and will? , 2006, Journal of personality.

[9]  A. Heathcote,et al.  Trust in Automation A Literature Review , 2018 .

[10]  Cynthia Breazeal,et al.  Social interactions in HRI: the robot view , 2004, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).

[11]  Abuse Humans and Automation : Use , Misuse , Disuse , , 2008 .

[12]  C. Nass,et al.  Machines and Mindlessness , 2000 .

[13]  E. E. Jones,et al.  The actor and the observer: Divergent perceptions of the causes of behavior. , 1972 .

[14]  John D. Lee,et al.  Review of a Pivotal Human Factors Article: “Humans and Automation: Use, Misuse, Disuse, Abuse” , 2008, Hum. Factors.

[15]  J. H. Davis,et al.  An Integrative Model Of Organizational Trust , 1995 .

[16]  John D. Lee,et al.  Trust in Automation: Designing for Appropriate Reliance , 2004 .

[17]  S. Asch Forming impressions of personality. , 1946, Journal of Abnormal Psychology.

[18]  M. Sieverding,et al.  Information About Low Participation in Cancer Screening Demotivates Other People , 2010, Psychological science.

[19]  F. Heider The psychology of interpersonal relations , 1958 .

[20]  Benjamin Kuipers,et al.  How can we trust a robot? , 2018, Commun. ACM.

[21]  Michael A. Olson,et al.  Implicit measures in social cognition. research: their meaning and use. , 2003, Annual review of psychology.

[22]  Peter A. Hancock,et al.  Can You Trust Your Robot? , 2011 .

[23]  E. Deci,et al.  Self-determination theory and the facilitation of intrinsic motivation, social development, and well-being. , 2000, The American psychologist.

[24]  Rosemarie E. Yagoda,et al.  You Want Me to Trust a ROBOT? The Development of a Human–Robot Interaction Trust Scale , 2012, International Journal of Social Robotics.

[25]  D. Lagnado,et al.  Judgments of cause and blame: The effects of intentionality and foreseeability , 2008, Cognition.

[26]  Raja Parasuraman,et al.  Humans and Automation: Use, Misuse, Disuse, Abuse , 1997, Hum. Factors.

[27]  Illah R. Nourbakhsh,et al.  A survey of socially interactive robots , 2003, Robotics Auton. Syst..

[28]  Anna Esposito,et al.  Introduction to the Special Issue “Beyond Industrial Robotics: Social Robots Entering Public and Domestic Spheres” , 2015, Inf. Soc..

[29]  Cynthia Breazeal,et al.  Social robots for health applications , 2011, 2011 Annual International Conference of the IEEE Engineering in Medicine and Biology Society.

[30]  Jessie Y. C. Chen,et al.  A Meta-Analysis of Factors Affecting Trust in Human-Robot Interaction , 2011, Hum. Factors.

[31]  Batya Friedman,et al.  “It's the computer's fault”: reasoning about computers as moral agents , 1995, CHI 95 Conference Companion.