An experimental study of public trust in AI chatbots in the public sector

Abstract This study investigates the public's initial trust in so-called “artificial intelligence” (AI) chatbots about to be introduced into use in the public sector. While the societal impacts of AI are widely speculated about, empirical testing remains rare. To narrow this gap, this study builds on theories of operators' trust in machines in industrial settings and proposes that initial public trust in chatbot responses depends on (i) the area of enquiry, since expectations about a chatbot's performance vary with the topic, and (ii) the purposes that governments communicate to the public for introducing the use of chatbots. Analyses based on an experimental online survey in Japan generated results indicating that, if a government were to announce its intention to use “AI” chatbots to answer public enquiries, the public's initial trust in their responses would be lower in the area of parental support than in the area of waste separation, with a moderate effect size. Communicating purposes that would directly benefit citizens, such as achieving uniformity in response quality and timeliness in responding, would enhance public trust in chatbots. Although the effect sizes are small, communicating these purposes might be still worthwhile, as it would be an inexpensive measure for a government to take.

[1]  Marijn Janssen,et al.  The challenges and limits of big data algorithms in technocratic governance , 2016, Gov. Inf. Q..

[2]  Regina A. Pomranky,et al.  The role of trust in automation reliance , 2003, Int. J. Hum. Comput. Stud..

[3]  J. G. Holmes,et al.  Trust in close relationships. , 1985 .

[4]  The Situational Context on the Nature of Political Philosophy , 2005 .

[5]  Lisa M. V. Gulick,et al.  Drivers and Consequences of Citizen Satisfaction: An Application of the American Customer Satisfaction Index Model to New York City , 2004 .

[6]  Helle Zinner Henriksen,et al.  Digital discretion: A systematic literature review of ICT and street-level discretion , 2018, Inf. Polity.

[7]  Asbjørn Følstad,et al.  Chatbots: changing user needs and motivations , 2018, Interactions.

[8]  Cees J. H. Midden,et al.  The effects of errors on system trust, self-confidence, and the allocation of control in route planning , 2003, Int. J. Hum. Comput. Stud..

[9]  Geert Bouckaert,et al.  Assessing the Relation Between Satisfaction with Public Service Delivery and Trust in Government. The Impact of the Predisposition of Citizens Toward Government on Evalutations of Its Performance , 2006 .

[10]  M. Coeckelbergh Can we trust robots? , 2011, Ethics and Information Technology.

[11]  N. Moray,et al.  Adaptive automation, trust, and self-confidence in fault management of time-critical tasks. , 2000 .

[12]  Robert Dale,et al.  The return of the chatbots , 2016, Natural Language Engineering.

[13]  Joshua B. Tenenbaum,et al.  Building machines that learn and think like people , 2016, Behavioral and Brain Sciences.

[14]  V. Weerakkody,et al.  Analysing the critical factors influencing trust in e-government adoption from citizens’ perspective: A systematic review and a conceptual framework , 2017 .

[15]  Martin Bichler,et al.  Responsible Data Science , 2017, Bus. Inf. Syst. Eng..

[16]  N Moray,et al.  Trust, control strategies and allocation of function in human-machine systems. , 1992, Ergonomics.

[17]  Lingling Gao,et al.  Examining the role of initial trust in user adoption of mobile payment services: an empirical investigation , 2015, Information Systems Frontiers.

[18]  S Lewandowsky,et al.  The dynamics of trust: comparing humans to automation. , 2000, Journal of experimental psychology. Applied.

[19]  Bonnie M. Muir,et al.  Trust Between Humans and Machines, and the Design of Decision Aids , 1987, Int. J. Man Mach. Stud..

[20]  Paulo Henrique de Souza Bermejo,et al.  How and where is artificial intelligence in the public sector going? A literature review and research agenda , 2019, Gov. Inf. Q..

[21]  D. Wiegmann,et al.  Similarities and differences between human–human and human–automation trust: an integrative review , 2007 .

[22]  D. Yates,et al.  Street-Level Bureaucracy: Dilemmas of the Individual in Public Services , 1981, Michigan Law Review.

[23]  Philip C. Treleaven,et al.  Algorithmic Government: Automating Public Services and Supporting Civil Servants in using Data Science Technologies , 2019, Comput. J..

[24]  F. C. Mosher,et al.  Democracy and the public service , 1968 .

[25]  Jan C. Weyerer,et al.  Artificial Intelligence and the Public Sector—Applications and Challenges , 2018, International Journal of Public Administration.

[26]  Eric Atwell,et al.  Chatbots: Are they Really Useful? , 2007, LDV Forum.

[27]  Bernard Barber,et al.  The Logic and Limits of Trust , 1983 .

[28]  Daniel R. Ilgen,et al.  Not All Trust Is Created Equal: Dispositional and History-Based Trust in Human-Automation Interactions , 2008, Hum. Factors.

[29]  Mariglynn Edlins,et al.  Ready to serve the public? The role of empathy in public service education programs , 2018 .

[30]  G. Allison Conceptual Models and the Cuban Missile Crisis , 1969, American Political Science Review.

[31]  Frédéric Kaplan,et al.  Who is Afraid of the Humanoid? Investigating Cultural Differences in the Acceptance of Robots , 2004, Int. J. Humanoid Robotics.

[32]  M. Bovens,et al.  From Street‐Level to System‐Level Bureaucracies: How Information and Communication Technology is Transforming Administrative Discretion and Constitutional Control , 2002 .

[33]  C. Lindblom THE SCIENCE OF MUDDLING THROUGH , 1959 .

[34]  W. Niskanen Bureaucracy and representative government , 1971 .

[35]  Clifford Nass,et al.  The media equation - how people treat computers, television, and new media like real people and places , 1996 .

[36]  Jeffrey B. Wenger,et al.  At the Discretion of Rogue Agents: How Automation Improves Women's Outcomes in Unemployment Insurance , 2008 .

[37]  B. Jack Copeland,et al.  Artificial Intelligence: A Philosophical Introduction , 1993 .

[38]  Rosario Espinal,et al.  Performance Still Matters , 2006 .

[39]  Francesca Rossi,et al.  AI4People—An Ethical Framework for a Good AI Society: Opportunities, Risks, Principles, and Recommendations , 2018, Minds and Machines.

[40]  Masooda Bashir,et al.  Trust in Automation , 2015, Hum. Factors.

[41]  Thomas J. Barth,et al.  Artificial Intelligence and Administrative Discretion , 1999 .

[42]  Ellen Enkel,et al.  Applied artificial intelligence and trust—The case of autonomous vehicles and medical assistance devices , 2016 .

[43]  Jaap J. Dijkstra,et al.  User agreement with incorrect expert system advice , 1999, Behav. Inf. Technol..

[44]  P. Agarwal,et al.  Public Administration Challenges in the World of AI and Bots , 2018, Public Administration Review.

[45]  John Kane,et al.  In Search of Prudence: The Hidden Problem of Managerial Reform , 2006 .

[46]  D. Hassabis,et al.  Neuroscience-Inspired Artificial Intelligence , 2017, Neuron.

[47]  John D. Lee,et al.  Trust in Automation: Designing for Appropriate Reliance , 2004 .