<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Wolfhart Totschnig[</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Fully autonomous AI</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Universidad Diego Portales</institution>
          ,
          <addr-line>Santiago</addr-line>
          ,
          <country country="CL">Chile</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>0000</year>
      </pub-date>
      <volume>0003</volume>
      <abstract>
        <p>Note: This is an extended abstract of a paper that has been accepted for publication in Science and Engineering Ethics. In the elds of arti cial intelligence and robotics, the term \autonomy" is generally used to mean the capacity of an arti cial agent to operate independently of human guidance. To create agents that are autonomous in this sense is the central aim of these elds. Until recently, the aim could be achieved only by restricting and controlling the conditions under which the agents will operate. The robots on an assembly line in a factory, for instance, perform their delicate tasks reliably because the surroundings have been meticulously prepared. Today, however, we are witnessing the creation of arti cial agents that are designed to function in \real-world"|that is, uncontrolled|environments. Self-driving cars, which are already in use, and \autonomous weapon systems," which are in development, are the most prominent examples. When such machines are called \autonomous," it is meant that they are able to choose by themselves, without human intervention, the appropriate course of action in the manifold situations they encounter.1 This way of using the term \autonomy" goes along with the assumption that the arti cial agent has a xed goal or \utility function," a set purpose with respect to which the appropriateness of its actions will be evaluated. So, in the rst example, the agent's purpose is to drive safely and e ciently from one place to another, and in the second example, it is to neutralize all and only enemy combatants in the chosen area of operation. It has thus been de ned and established, in general terms, what the agent is supposed to do. The attribute \autonomous" concerns only whether the agent will be able to carry out the given general instructions in concrete situations. From a philosophical perspective, this notion of autonomy seems oddly weak. For, in philosophy, the term is generally used to refer to a stronger capacity, namely the capacity, as Kant put it, to \give oneself the law" (Kant [1785] 1998, 4:440{441), to decide by oneself what one's goal or principle of action will be. This understanding of the term derives from its Greek etymology (auto = \by oneself," nomos = \law"). An instance of such autonomy would be an agent who decides, by itself, to devote its e orts to a certain project|the attainment of 1 For prominent instances of this usage, see Russell &amp; Norvig's popular textbook Arti cial intelligence: A modern approach (2010, 18), Anderson &amp; Anderson's introduction to their edited volume Machine ethics (2011, 1), the papers collected in the volume Autonomy and arti cial intelligence (Lawless et al. 2017) and especially the one by Tessier (2017), as well as Muller (2012), Mindell (2015, ch. 1), and Johnson &amp; Verdicchio (2017).</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>W. Totschnig
knowledge, say, or the realization of justice. In contrast, any agent that has a
predetermined and immutable goal or purpose would not be considered autonomous
in this sense.</p>
      <p>The aim of the present paper is to argue that an arti cial agent can possess
autonomy as understood in philosophy|or \full autonomy," as I will call it for
short. \Can" is here intended in the sense of general possibility, not in the sense
of current feasibility. I contend that the possibility of a fully autonomous AI
cannot be excluded, but do not mean to imply that such an AI can be created
today.</p>
      <p>
        My argument stands in opposition to the predominant view in the literature
on the long-term prospects and risks of arti cial intelligence. The predominant
view is that an arti cial agent cannot exhibit full autonomy because it cannot
rationally change its own nal goal, since changing the nal goal is
counterproductive with respect to that goal and hence undesirable
        <xref ref-type="bibr" rid="ref10 ref11 ref12 ref15 ref16 ref17 ref18 ref19 ref2 ref20 ref3 ref4">(Yudkowsky 2001, 2008,
2011, 2012; Bostrom 2002, 2014; Omohundro 2008, 2012, 2016; Yampolskiy &amp;
Fox 2012, 2013; Domingos 2015)</xref>
        . I will challenge this view by showing that it is
based on questionable assumptions about the nature of goals and values. I will
argue that a general arti cial intelligence|i.e., an arti cial intelligence that, like
human beings, develops a general understanding of the world and of itself|may
very well come to change its nal goal in the course of its development.
      </p>
      <p>This issue is obviously of great importance for how we are to assess the
longterm prospects and risks of arti cial intelligence. If arti cial agents can reach full
autonomy, which law will they give themselves when that happens? In particular,
what con dence can we have that the chosen law will include respect for human
beings?</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Anderson</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Anderson</surname>
          </string-name>
          , S.L.:
          <article-title>General introduction</article-title>
          . In: Anderson,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Anderson</surname>
          </string-name>
          ,
          <string-name>
            <surname>S.L</surname>
          </string-name>
          . (eds.) Machine ethics, pp.
          <volume>1</volume>
          {
          <fpage>4</fpage>
          . Cambridge University Press (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Bostrom</surname>
          </string-name>
          , N.:
          <article-title>Existential risks</article-title>
          .
          <source>Journal of Evolution and Technology</source>
          <volume>9</volume>
          (
          <issue>1</issue>
          ) (
          <year>2002</year>
          ), http://www.jetpress.org/volume9/risks.html
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Bostrom</surname>
          </string-name>
          , N.:
          <string-name>
            <surname>Superintelligence</surname>
          </string-name>
          . Oxford University Press (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Domingos</surname>
            ,
            <given-names>P.:</given-names>
          </string-name>
          <article-title>The master algorithm</article-title>
          .
          <source>Basic Books</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Johnson</surname>
          </string-name>
          , D.G.,
          <string-name>
            <surname>Verdicchio</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Reframing AI discourse</article-title>
          .
          <source>Minds and Machines</source>
          <volume>27</volume>
          (
          <issue>4</issue>
          ),
          <volume>575</volume>
          {
          <fpage>590</fpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Kant</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Groundwork of the metaphysics of morals</article-title>
          . Cambridge Texts in the History of Philosophy, Cambridge University Press (
          <year>1998</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Lawless</surname>
            ,
            <given-names>W.F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mittu</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sofge</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Russell</surname>
          </string-name>
          , S. (eds.):
          <article-title>Autonomy and arti cial intelligence</article-title>
          . Springer International Publishing (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Mindell</surname>
            ,
            <given-names>D.A.</given-names>
          </string-name>
          :
          <article-title>Our robots, ourselves</article-title>
          .
          <source>Viking</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9. Muller, V.C.:
          <article-title>Autonomous cognitive systems in real-world environments</article-title>
          .
          <source>Cognitive Computation</source>
          <volume>4</volume>
          (
          <issue>3</issue>
          ),
          <volume>212</volume>
          {
          <fpage>215</fpage>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Omohundro</surname>
            ,
            <given-names>S.M.:</given-names>
          </string-name>
          <article-title>The nature of self-improving arti cial intelligence (</article-title>
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Omohundro</surname>
            ,
            <given-names>S.M.:</given-names>
          </string-name>
          <article-title>Rational arti cial intelligence for the greater good</article-title>
          . In: Eden,
          <string-name>
            <given-names>A.H.</given-names>
            ,
            <surname>Moor</surname>
          </string-name>
          ,
          <string-name>
            <surname>J.H.,</surname>
          </string-name>
          <article-title>S raker</article-title>
          ,
          <string-name>
            <given-names>J.H.</given-names>
            ,
            <surname>Steinhart</surname>
          </string-name>
          , E. (eds.) Singularity hypotheses, pp.
          <volume>161</volume>
          {
          <fpage>176</fpage>
          . Springer (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Omohundro</surname>
            ,
            <given-names>S.M.:</given-names>
          </string-name>
          <article-title>Autonomous technology and the greater human good</article-title>
          . In: Muller, V.C. (ed.)
          <source>Risks of arti cial intelligence</source>
          , pp.
          <volume>9</volume>
          {
          <fpage>27</fpage>
          . CRC Press (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Russell</surname>
            ,
            <given-names>S.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Norvig</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Arti cial intelligence</article-title>
          .
          <source>Prentice Hall</source>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Tessier</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Robots autonomy</article-title>
          . In: Lawless,
          <string-name>
            <given-names>W.F.</given-names>
            ,
            <surname>Mittu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Sofge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            ,
            <surname>Russell</surname>
          </string-name>
          , S. (eds.)
          <article-title>Autonomy and arti cial intelligence</article-title>
          , pp.
          <volume>179</volume>
          {
          <fpage>194</fpage>
          . Springer International Publishing (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Yampolskiy</surname>
            ,
            <given-names>R.V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fox</surname>
          </string-name>
          , J.:
          <article-title>Arti cial general intelligence and the human mental model</article-title>
          . In: Eden,
          <string-name>
            <given-names>A.H.</given-names>
            ,
            <surname>Moor</surname>
          </string-name>
          ,
          <string-name>
            <surname>J.H.,</surname>
          </string-name>
          <article-title>S raker</article-title>
          ,
          <string-name>
            <given-names>J.H.</given-names>
            ,
            <surname>Steinhart</surname>
          </string-name>
          , E. (eds.) Singularity hypotheses, pp.
          <volume>129</volume>
          {
          <fpage>145</fpage>
          . Springer (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Yampolskiy</surname>
            ,
            <given-names>R.V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fox</surname>
          </string-name>
          , J.:
          <article-title>Safety engineering for arti cial general intelligence</article-title>
          .
          <source>Topoi</source>
          <volume>32</volume>
          (
          <issue>2</issue>
          ),
          <volume>217</volume>
          {
          <fpage>226</fpage>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Yudkowsky</surname>
          </string-name>
          , E.:
          <article-title>Creating friendly AI 1.0. The Singularity Institute (</article-title>
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Yudkowsky</surname>
          </string-name>
          , E.:
          <article-title>Arti cial intelligence as a positive and negative factor in global risk</article-title>
          . In: Bostrom,
          <string-name>
            <given-names>N.</given-names>
            ,
            <surname>Cirkovic</surname>
          </string-name>
          , M.M. (eds.) Global catastrophic risks, pp.
          <volume>308</volume>
          {
          <fpage>345</fpage>
          . Oxford University Press (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Yudkowsky</surname>
          </string-name>
          , E.:
          <article-title>Complex value systems in friendly AI</article-title>
          . In: Schmidhuber,
          <string-name>
            <given-names>J.</given-names>
            ,
            <surname>Thrisson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.R.</given-names>
            ,
            <surname>Looks</surname>
          </string-name>
          , M. (eds.) Arti cial general intelligence, pp.
          <volume>388</volume>
          {
          <fpage>393</fpage>
          . Springer (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Yudkowsky</surname>
          </string-name>
          , E.:
          <article-title>Friendly arti cial intelligence</article-title>
          . In: Eden,
          <string-name>
            <given-names>A.H.</given-names>
            ,
            <surname>Moor</surname>
          </string-name>
          ,
          <string-name>
            <surname>J.H.,</surname>
          </string-name>
          <article-title>S raker</article-title>
          ,
          <string-name>
            <given-names>J.H.</given-names>
            ,
            <surname>Steinhart</surname>
          </string-name>
          , E. (eds.) Singularity hypotheses, pp.
          <volume>181</volume>
          {
          <fpage>193</fpage>
          . Springer (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>