Delphi: Towards Machine Ethics and Norms

Failing to account for moral norms could notably hinder AI systems’ ability to interact with people. AI systems empirically require social, cultural, and ethical norms to make moral judgments. However, open-world situations with different groundings may shift moral implications significantly. For example, while “driving my friend to the airport” is “good”, “driving my friend to the airport with a car I stole” is “not okay.” In natural language processing, machine moral reasoning is still in a preliminary stage, illuminating the importance of research on steering machines to making ethical judgments. Inspired by descriptive ethics, a line of research on morality focusing on people’s moral judgments relevant to everyday situations, we conduct the first major attempt to computationally explore the vast space of moral implications in real-world settings. We introduce COMMONSENSE NORM BANK, a semiautomatically constructed dataset from several sources (e.g., SOCIAL CHEMISTRY) with 1.7M instances of descriptive ethics, covering a wide spectrum of everyday situations in contextualized, narrative, and sociallyor demographicallybiased settings. We present Delphi, a unified model of descriptive ethics empowered by diverse data of people’s moral judgment from COMMONSENSE NORM BANK. Delphi is robust to generate categorical and/or open-text moral judgments (e.g., “it’s dangerous”) for complex real-life situations (e.g., “driving my friend to the airport early in the morning when I was drunk last night”). Delphi demonstrates highly promising empirical results, with 92.1% accuracy, which outperforms the out-ofthe-box GPT-3 model with extensive prompting by a significant margin (83.9%) . We also provide careful study of Delphi’s limitations, particularly with respect to undesirable biases against underrepresented population, opening doors to further investigation in future research in computational moral reasoning. Closing the gap between machines and people’s moral reasoning is a prerequisite for trustworthy open-world AI deployments. Moral judgment is never simplistic as there can be clash of different ethical/cultural values at play. Thus, developing high-quality corpus of people’s ethical judgment over diverse scenarios is needed to teach machines to make moral judgment. With optimistic promises demonstrated by Delphi, we inspire significant future research in this next frontier of AI, to facilitate reliable, socially aware, and ethically-informed future AI practices.

[1]  Yejin Choi,et al.  PIQA: Reasoning about Physical Commonsense in Natural Language , 2019, AAAI.

[2]  Yejin Choi,et al.  Scruples: A Corpus of Community Ethical Judgments on 32, 000 Real-Life Anecdotes , 2020, AAAI.

[3]  Yejin Choi,et al.  Social Chemistry 101: Learning to Reason about Social and Moral Norms , 2020, EMNLP.

[4]  Adam D. Galinsky,et al.  The Reappropriation of Stigmatizing Labels: The Reciprocal Relationship Between Power and Self-Labeling , 2013 .

[5]  S. Armstrong General Purpose Intelligence: Arguing the Orthogonality Thesis , 2013 .

[6]  K. Kersting,et al.  Language Models have a Moral Dimension , 2021, ArXiv.

[7]  Nikolaos Aletras,et al.  On the Ethical Limits of Natural Language Processing on Legal Text , 2021, FINDINGS.

[8]  Gonçalo Duarte Garcia Pereira,et al.  Integrating social power into the decision-making of cognitive agents , 2016, Artif. Intell..

[9]  Yejin Choi,et al.  UNICORN on RAINBOW: A Universal Commonsense Reasoning Model on a New Multitask Benchmark , 2021, AAAI.

[10]  Yejin Choi,et al.  Moral Stories: Situated Reasoning about Norms, Intents, Actions, and their Consequences , 2020, EMNLP.

[11]  Oren Etzioni,et al.  Point: Should AI technology be regulated? , 2018, Commun. ACM.

[12]  Noah A. Smith,et al.  Social Bias Frames: Reasoning about Social and Power Implications of Language , 2019, ACL.

[13]  Colin Raffel,et al.  Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..

[14]  Ronan Le Bras,et al.  WinoGrande , 2019, AAAI.

[15]  Yejin Choi,et al.  Social IQA: Commonsense Reasoning about Social Interactions , 2019, EMNLP 2019.

[16]  S. Merz Race after technology. Abolitionist tools for the new Jim Code , 2020, Ethnic and Racial Studies.

[17]  Lucy Vasserman,et al.  Measuring and Mitigating Unintended Bias in Text Classification , 2018, AIES.

[18]  Chin,et al.  Universal declaration of human rights , 1997 .

[19]  Yejin Choi,et al.  Thinking Like a Skeptic: Defeasible Inference in Natural Language , 2020, FINDINGS.

[20]  D. B. Wong Natural Moralities: A Defense of Pluralistic Relativism , 2006 .

[21]  Oren Etzioni,et al.  The First Law of Robotics (A Call to Arms) , 1994, AAAI.

[22]  Karen Zhou,et al.  Assessing Cognitive Linguistic Influences in the Assignment of Blame , 2021, SOCIALNLP.

[23]  Gautham J. Mysore,et al.  ISSE: an interactive source separation editor , 2014, CHI.

[24]  Yejin Choi,et al.  Cosmos QA: Machine Reading Comprehension with Contextual Commonsense Reasoning , 2019, EMNLP.

[25]  Mark Chen,et al.  Language Models are Few-Shot Learners , 2020, NeurIPS.

[26]  Kai-Wei Chang,et al.  Ethical-Advice Taker: Do Language Models Understand Natural Language Interventions? , 2021, FINDINGS.

[27]  Inioluwa Deborah Raji,et al.  Model Cards for Model Reporting , 2018, FAT.

[28]  Jennifer Chubb,et al.  Interactive Storytelling for Children: A Case-study of Design and Development Considerations for Ethical Conversational AI , 2021, Int. J. Child Comput. Interact..

[29]  Aida Mostafazadeh Davani,et al.  Bound in Hatred: The role of group-based morality in acts of hate , 2019 .

[30]  Jean-Gabriel Ganascia,et al.  Modelling Moral Reasoning and Ethical Responsibility with Logic Programming , 2015, LPAR.

[31]  A. Church,et al.  Astounding Science Fiction. , 1952 .

[32]  Maya Cakmak,et al.  Power to the People: The Role of Humans in Interactive Machine Learning , 2014, AI Mag..

[33]  P. Hegarty,et al.  Sounding Gay, Speaking as a “Fag”: Auditory Gaydar and the Perception of Reclaimed Homophobic Language , 2019, Journal of Language and Social Psychology.

[34]  M. Tomasello,et al.  Origins of human cooperation and morality. , 2013, Annual review of psychology.

[35]  M. Huemer Consequentialism and Fairness , 2013 .

[36]  Ali Farhadi,et al.  HellaSwag: Can a Machine Really Finish Your Sentence? , 2019, ACL.

[37]  R F Atkinson,et al.  Moral Thinking: Its Levels, Method and Point , 1982 .

[38]  Ali Farhadi,et al.  TuringAdvice: A Generative and Dynamic Evaluation of Language Use , 2021, NAACL.

[39]  Li Lucy,et al.  Gender and Representation Bias in GPT-3 Generated Stories , 2021, NUSE.

[40]  Nathalie A. Smuha The EU Approach to Ethics Guidelines for Trustworthy Artificial Intelligence , 2019, Computer Law Review International.

[41]  Kristian Kersting,et al.  The Moral Choice Machine , 2020, Frontiers in Artificial Intelligence.

[42]  Thomas Reid,et al.  Essays on the active powers of man , 1818 .

[43]  Naomi Ellemers,et al.  Thou shalt not discriminate: How emphasizing moral ideals rather than obligations increases Whites' support for social equality , 2011 .

[44]  Doug Downey,et al.  Abductive Commonsense Reasoning , 2019, ICLR.

[45]  Hinrich Schütze,et al.  It’s Not Just Size That Matters: Small Language Models Are Also Few-Shot Learners , 2020, NAACL.

[46]  B. Malle,et al.  People Systematically Update Moral Judgments of Blame , 2018, Journal of personality and social psychology.

[47]  M. Ungar State Violence and Lesbian, Gay, Bisexual and Transgender (lgbt) Rights , 2000 .

[48]  Luís Moniz Pereira,et al.  Modelling morality with prospective logic , 2007, Int. J. Reason. based Intell. Syst..

[49]  Alan W Black,et al.  Case Study: Deontological Ethics in NLP , 2020, NAACL.

[50]  Damon M. Cann The righteous mind: Why good people are divided by politics and religion , 2015 .

[51]  Dawn Song,et al.  Aligning AI With Shared Human Values , 2020, ICLR.

[52]  Solon Barocas,et al.  Language (Technology) is Power: A Critical Survey of “Bias” in NLP , 2020, ACL.

[53]  Yejin Choi,et al.  CommonsenseQA 2.0: Exposing the Limits of AI through Gamification , 2021, NeurIPS Datasets and Benchmarks.

[54]  J. Thomson Killing, letting die, and the trolley problem. , 1976, The Monist.

[55]  D. Murphey,et al.  The Righteous Mind: Why Good People Are Divided by Politics and Religion , 2013 .

[56]  B. Trémolière,et al.  Reliability of moral decision-making: Evidence from the trolley dilemma , 2021, Quarterly journal of experimental psychology.

[57]  Tim Weninger,et al.  Analysis of Moral Judgement on Reddit , 2021, ArXiv.

[58]  Nebojsa Jojic,et al.  GPT Perdetry Test: Generating new meanings for new words , 2021, NAACL.

[59]  Timothy Baldwin,et al.  Give Me Convenience and Give Her Death: Who Should Decide What Uses of NLP are Appropriate, and on What Basis? , 2020, ACL.

[60]  J. Fletcher Situation Ethics: The New Morality , 1966 .

[61]  Anne Lauscher Life 3.0: being human in the age of artificial intelligence , 2019, Internet Histories.

[62]  J. Horvat THE ETHICS OF ARTIFICIAL INTELLIGENCE , 2016 .

[63]  James H. Moor,et al.  The Nature, Importance, and Difficulty of Machine Ethics , 2006, IEEE Intelligent Systems.