Fully autonomous AI
Keywords: artificial intelligence, autonomy, goals, normativity
Abstract
In the fields of artificial intelligence and robotics, the term âautonomyâ is generally used to mean the capacity of an artificial agent to operate independently of human guidance. It is thereby assumed that the agent has a fixed goal or âutility functionâ with respect to which the appropriateness of its actions will be evaluated. From a philosophical perspective, this notion of autonomy seems oddly weak. For, in philosophy, the term is generally used to refer to a stronger capacity, namely the capacity to âgive oneself the law,â to decide by oneself what oneâs goal or principle of action will be. The predominant view in the literature on the long-term prospects and risks of artificial intelligence is that an artificial agent cannot exhibit such autonomy because it cannot rationally change its own final goal, since changing the final goal is counterproductive with respect to that goal and hence undesirable. The aim of this paper is to challenge this view by showing that it is based on questionable assumptions about the nature of goals and values. I argue that a general AI may very well come to modify its final goal in the course of developing its understanding of the world. This has important implications for how we are to assess the long-term prospects and risks of artificial intelligence.
Más información
| Título de la Revista: | Science and Engineering Ethics |
| Volumen: | 26 |
| Número: | 5 |
| Editorial: | Springer Science and Business Media B.V. |
| Fecha de publicación: | 2020 |
| Página final: | 2485 |
| Idioma: | English |
| DOI: |
10.1007/s11948-020-00243-z |
| Notas: | Web of Science Core Collection: Science Citation Index Expanded | Social Sciences Citation Index |