Fully autonomous AI
Keywords: artificial intelligence, autonomy, goals, normativity
Abstract
In the fields of artificial intelligence and robotics, the term “autonomy” is generally used to mean the capacity of an artificial agent to operate independently of human guidance. It is thereby assumed that the agent has a fixed goal or “utility function” with respect to which the appropriateness of its actions will be evaluated. From a philosophical perspective, this notion of autonomy seems oddly weak. For, in philosophy, the term is generally used to refer to a stronger capacity, namely the capacity to “give oneself the law,” to decide by oneself what one’s goal or principle of action will be. The predominant view in the literature on the long-term prospects and risks of artificial intelligence is that an artificial agent cannot exhibit such autonomy because it cannot rationally change its own final goal, since changing the final goal is counterproductive with respect to that goal and hence undesirable. The aim of this paper is to challenge this view by showing that it is based on questionable assumptions about the nature of goals and values. I argue that a general AI may very well come to modify its final goal in the course of developing its understanding of the world. This has important implications for how we are to assess the long-term prospects and risks of artificial intelligence.
Más información
Título de la Revista: | SCIENCE AND ENGINEERING ETHICS |
Editorial: | Springer |
Fecha de publicación: | 2020 |
Idioma: | inglés |
DOI: |
10.1007/s11948-020-00243-z |
Notas: | Web of Science Core Collection: Science Citation Index Expanded | Social Sciences Citation Index |