Ethical Issues with Artificial Ethics Assistants

Abstract

This chapter examines the possibility of using artificial intelligence (AI) technologies to improve human moral reasoning and decision-making. The authors characterize such technologies as artificial ethics assistants (AEAs). The authors focus on just one part of the AI-aided moral improvement question: the case of the individual who wants to improve their morality, where what constitutes an improvement is evaluated by the individual’s own values. The authors distinguish three broad areas in which an individual might think their own moral reasoning and decision-making could be improved: one’s actions, character, or other attributes fall short of one’s values and moral beliefs; one sometimes misjudges or is uncertain about what the right thing to do is, given one’s values; or one is uncertain about some fundamental moral questions or recognizes a possibility that some of one’s core moral beliefs and values are mistaken. The authors sketch why one might think AI tools could be used to support moral improvement in those areas and distinguish two types of assistance: preparatory assistance, including advice and training supplied in advance of moral deliberation, and on-the-spot assistance, including on-the-spot advice and facilitation of moral functioning over the course of moral deliberation. Then, the authors turn to ethical issues that AEAs might raise, looking in particular at three under-appreciated problems posed by the use of AI for moral self-improvement: namely, reliance on sensitive moral data, the inescapability of outside influences on AEAs, and AEA usage prompting the user to adopt beliefs and make decisions without adequate reasons.

Publication
The Oxford Handbook of Digital Ethics
Elizabeth O'Neill
Elizabeth O'Neill
Assistant Professor