<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Four Preconditions for Solving MC4 Machine Consciousness</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>David Gamez</string-name>
          <email>d.gamez@mdx.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, Middlesex University</institution>
          ,
          <addr-line>London, NW4 4BT</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>A machine is MC4 conscious if it has phenomenal experiences that are comparable to human conscious experiences. From an ethical point of view it is important to know whether we have created MC4 consciousness in a machine. MC4 consciousness research can also contribute to the development of general theories of human consciousness. This paper discusses four problems that have to be solved before we will be able to address MC4 machine consciousness in a systematic way: We need more clarity about the measurement of consciousness, we need better ways of describing the physical world and consciousness, and we need to reach agreement about the final form that a theory of consciousness should take. When these problems have been addressed we will be able to develop scientific theories of consciousness that can make accurate believable predictions about MC4 consciousness in machines.</p>
      </abstract>
      <kwd-group>
        <kwd />
        <kwd>consciousness</kwd>
        <kwd>machine consciousness</kwd>
        <kwd>artificial consciousness</kwd>
        <kwd>science of consciousness</kwd>
        <kwd>neural correlates</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        It is often helpful to distinguish four types of machine consciousness [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]:
 MC1. Machines with the same external behavior as conscious systems. Humans
behave in particular ways when they are conscious. For example, they are alert,
they can respond to novel situations, they can inwardly execute sequences of
problem-solving steps and they can learn. AI systems can exhibit some or all of these
external behaviors.
 MC2. Models of the correlates of consciousness. Theories about the neural and
functional correlates of consciousness in humans can be modeled in a computer.
 MC3. Models of consciousness. Phenomenal experiences have characteristic
features, which can be modeled in computers and used to control robots.
 MC4. Machines that are phenomenally conscious. When humans are conscious
they are immersed in a world of colors, smells, sounds, etc. These are not
properties of the physical world – they are constituents of conscious experience. A
machine that was immersed in a world of colors, smells and sounds would be MC4
conscious.
      </p>
      <p>These different types of machine consciousness can be combined. For example, a
robot that displayed conscious external behavior (MC1) could be controlled by a
model of consciousness (MC3) and have phenomenal experiences (MC4).</p>
      <p>This paper will address the question of MC4 consciousness: artificial systems that
have phenomenal experiences that are similar to our spatially and temporally
distributed experiences of color, smell, taste, etc. This type of machine consciousness is
ethically significant because MC4 conscious machines could suffer. Research on
MC4 machine consciousness could also be a key factor in the discovery of general
theories of consciousness that are not limited to the neurons inside the human
biological machine.</p>
      <p>MC4 machine consciousness can be solved when we have developed scientific
theories of consciousness that can make accurate detailed predictions about
consciousness in any physical system. This will only become possible after four
preconditions have been met. First, we need to reach agreement about how consciousness
can be measured and abandon the idea that consciousness can be measured through
machines’ external behavior (Section 2). Next we need to move away from neural
correlates of consciousness and find new ways of describing spatiotemporal physical
patterns in the brain that could form the basis for generalizable theories of
consciousness (Section 3). Third, we need less anthropomorphic ways of describing
consciousness (Section 4). Finally we have to drop our desire for intuitively satisfying
explanations of consciousness and search for mathematical relationships between formal
descriptions of consciousness and formal descriptions of the physical world (Section
5). When these preconditions have been met we will be able to use human and animal
experiments to discover mathematical theories of consciousness that can be
generalized to make believable predictions about MC4 consciousness in artificial systems.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Measurement of Consciousness</title>
      <p>We cannot directly measure other people’s consciousness: we have to measure
consciousness through external behavior (first-person reports). An inference from
external behavior to consciousness only works when the external behavior is generated by
a system that we believe is capable of consciousness – for example, when conscious
external behavior is exhibited by normal humans and primates. The inference from
external behavior to consciousness is less certain with brain-damaged patients, infants
and cephalopods.</p>
      <p>
        Some people believe that conscious external behavior can be used to infer the
presence of consciousness in artificial systems. If this was the case, the problem of MC4
machine consciousness would be solved. Any machine that exhibited conscious
external behavior would be judged to be phenomenally conscious and we could easily
identify the correlates of consciousness in machines. The problem with this position is
that it is extremely easy to write a computer program that mimics human first-person
reports about consciousness. For example: “cout&lt;&lt;’I am conscious of a red apple.’;”
This behavior can be produced by a giant lookup table [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and it is possible to
interpret any sequence of physical states as the execution of this program [
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ], including
a sequence of states in the brain of someone who is not conscious. When humans
execute this program in their heads they are not conscious of the red apple [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. These
problems do not go away when the simple one line program is replaced by a super
intelligent AI that passes the Turing Test, and they are also present when part of a
person’s brain is replaced with a functionally equivalent silicon chip [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>These problems prevent us from using external behavior to measure consciousness
in artificial systems, regardless of how humanlike that behavior is. If we cannot
measure consciousness in artificial systems, we cannot do MC4 consciousness
experiments on artificial systems. Instead, we will have to discover the relationship
between consciousness and the physical world in humans (and similar systems) and then
generalize these results to animals and artificial systems. To make this generalization
possible we have to rethink the way in which we study consciousness in humans. We
need to describe the physical states of human brains in a way that will enable us to
apply our theories of consciousness to non-biological systems. We also have to
describe consciousness without the anthropocentrism of natural language. These
challenges are covered in the next two sections.
3</p>
    </sec>
    <sec id="sec-3">
      <title>Description of the Physical World</title>
      <p>
        The measurement problem forces us to develop our theories of consciousness through
experiments on humans and similar animals. Although there have been promising
results on the neural correlates of consciousness [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], it is difficult to generalize them
to animals that have different brain architectures and neurons, and it is impossible to
apply them to artificial systems that are controlled by synthetic or silicon neurons.
      </p>
      <p>
        It might be thought that the generalization problem could be solved by linking
consciousness to functions, computations or information patterns in the brain. If
consciousness is linked to computations in the human brain, then a robot that executed
the same computations would also be conscious. A key problem with functional,
computational and informational theories of consciousness is that functions,
computations and information are not objective properties of the physical world. Information
appears when you apply a human-defined interface to the physical world [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] and
different interfaces lead to different information sets. Functions are equivalent to
computations, which are subjective uses that humans make of physical objects, not objective
properties of physical objects [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The physical universe does not contain functions,
computations or information. The subjectivity of functions, computations and
information makes it impossible to prove that they are linked to consciousness in the
human brain. It is also possible to interpret any sequence of physical states as a
particular information pattern or computation, which leads to contradictory predictions about
the consciousness that is associated with a sequence of physical states [
        <xref ref-type="bibr" rid="ref1 ref8 ref9">1, 8, 9</xref>
        ].
      </p>
      <p>An objective scientific approach to consciousness has to look for connections
between spatiotemporal physical patterns and consciousness. Much of the recent
research on consciousness has focused on the relationship between neural activity
patterns and conscious states. However, this will not lead to a theory that can be
generalized to artificial systems. Neurons are defined in a specific biological context – the
brains of animals – we have no formal definition that would enable us to
unambiguously identify neurons outside of this context. Suppose we genetically engineer a
sequence of 100 hybrids between neurons and liver cells: the first cell is 100 %
neuron; the middle cell is 50% neuron; the last cell is 100% liver. We have no systematic
way of classifying intermediate cells in this sequence. Or suppose we synthesize an
approximate neuron from basic biological components – we have no idea whether
synthetic or silicon neurons could form correlates of consciousness. A generalizable
theory of consciousness has to be based on precisely defined spatiotemporal structures
that can be unambiguously identified in any physical system. We will have to define
neurons more precisely if we want to generalize the results from the neural correlates
of consciousness, or we could base our theories of consciousness on other properties
of the physical world, such as electromagnetic waves.
4</p>
    </sec>
    <sec id="sec-4">
      <title>Description of Consciousness</title>
      <p>The majority of work on consciousness has been based on contrastive analyses of
conscious and unconscious brains. This gives us valuable information about the
neural activity patterns that are necessary for any conscious state to occur – neural
activity patterns that are common to all conscious states. There has been less work on the
detailed relationships between the contents of consciousness and physical states.</p>
      <p>
        Conscious states are typically described in natural language. But generalizable
theories of consciousness cannot be based on natural language because it is vague, highly
compressed and context dependent. It is likely that artificial systems will have
radically different experiences that will be impossible to describe in human language.
Natural language descriptions of consciousness are also likely to be highly misleading
because artificial systems might have different representations of the temporal and
spatial properties of objects [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ].
      </p>
      <p>
        In some experiments conscious contents are specified using the stimuli that
produced the conscious experiences. For example, in the work on brain reading with
fMRI the decoded conscious experiences are presented as videos, which the subject
compares with their own conscious experiences [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. This approach works with
humans because most humans have roughly the same conscious experience when they
are exposed to the same stimulus. But the conscious experience that I have when I
view a video is extremely unlikely to be the same as the consciousness of an artificial
system that views the same video, so the video cannot be used to describe the
consciousness of the artificial system.
      </p>
      <p>
        To address this problem we need to find new less anthropocentric ways of
describing consciousness that will enable us to generalize the experimental results on human
consciousness to artificial systems.1
1 One solution to this problem has been put forward by Balduzzi and Tononi, who suggested
how states of consciousness could be described using high dimensional mathematical
structures [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
      </p>
    </sec>
    <sec id="sec-5">
      <title>Theories of Consciousness</title>
      <p>There is a lack of general agreement about the final form that theories of
consciousness should take. A generalizable theory of consciousness should meet the following
criteria:
1. Can generate testable predictions.
2. Is applicable to any physical system.
3. Compact (Occam’s razor).
4. Based on objective properties of the physical world.</p>
      <p>Functional, computational and informational theories of consciousness meet criteria
1-3. However, as explained in Section 3, they fail to meet Criteria 4 because they are
not based on objective properties of the physical world. Theories about the neural
correlates of consciousness meet Criteria 4, but they are often not compact or
generalizable, and they have a weak ability to generate testable predictions.</p>
      <p>
        Many people are looking for an intuitively satisfying explanation of the
relationship between consciousness and the physical world. They want to make an
imaginative transition between a physical state (for example a mental image of a physical
brain) and a conscious state (for example, the color red). We are never going to get
this type of theory because we can only imagine conscious experiences, not the
invisible physical world as it is in itself without any conscious properties. In principle we
might be able to discover a theory that will enable us to make an imaginative
transition from a conscious experience of a brain to another conscious experience. But this
would only become intuitively convincing once we had learnt which brain patterns
are linked to conscious experiences [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The limitations of brain scanning technology
and human memory are likely to rule this out for the foreseeable future.
      </p>
      <p>
        The most plausible type of theory that meets all of the criteria is a mathematical
relationship between a formal description of the physical world (see Section 3) and a
formal description of consciousness (see Section 4). Compact mathematical theories
are the gold standard in many sciences, they can generate testable predictions and
they can be applied to any physical system. The most impressive mathematical theory
of consciousness that has been developed so far is Tononi’s information integration
theory [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. Although it has serious limitations, it does show how we might be able to
develop workable mathematical theories of consciousness in the future.
      </p>
      <p>To develop fine grained mathematical theories of consciousness we need high
resolution data from the brain. However, the vast amount of data that could
potentially be recorded from a brain could not be comprehended by a single human brain. This
suggests that humans are not likely to be capable of discovering mathematical
theories of consciousness. We could address this problem by using artificial intelligence to
search for mathematical relationships between consciousness and the physical world.</p>
    </sec>
    <sec id="sec-6">
      <title>Conclusion</title>
      <p>This paper has discussed four problems that have to be addressed before we can
develop a scientific solution to MC4 machine consciousness. We need to reach
agreement about how consciousness can be measured and abandon the idea that
consciousness can be inferred from a machine’s external behavior. Scientific work on
consciousness needs to move away from neural correlates and start to look for more
accurately defined spatiotemporal structures in the human brain that will lead to
generalizable theories of consciousness. We need to develop ways of describing
consciousness that avoid the anthropocentrism and context dependence of natural language. We
have to drop our desire for intuitively satisfying explanations of consciousness and
search for mathematical relationships between formal descriptions of consciousness
and formal descriptions of the physical world. If these challenges can be addressed,
we will have made significant progress towards tackling MC4 machine consciousness
in a scientifically plausible way.
7</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Gamez</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Human and Machine Consciousness</article-title>
          . Open Book Publishers, Cambridge (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Block</surname>
          </string-name>
          , N.:
          <article-title>Troubles with Functionalism</article-title>
          . In: Eckert, M. (ed.)
          <source>Theories of Mind: An Introductory Reader</source>
          , pp.
          <fpage>97</fpage>
          -
          <lpage>102</lpage>
          . Rowman &amp; Littlefield,
          <string-name>
            <surname>Maryland</surname>
          </string-name>
          (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Putnam</surname>
          </string-name>
          , H.:
          <article-title>Representation and Reality</article-title>
          . MIT Press, Cambridge, Massachusetts; London (
          <year>1988</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Bishop</surname>
            ,
            <given-names>J.M.</given-names>
          </string-name>
          :
          <string-name>
            <given-names>A Cognitive</given-names>
            <surname>Computation</surname>
          </string-name>
          <article-title>Fallacy? Cognition, Computations and Panpsychism</article-title>
          .
          <source>Cognitive Computation</source>
          <volume>1</volume>
          ,
          <fpage>221</fpage>
          -
          <lpage>233</lpage>
          (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Searle</surname>
            ,
            <given-names>J.R.</given-names>
          </string-name>
          : Minds, Brains, and
          <string-name>
            <surname>Programs</surname>
          </string-name>
          .
          <source>Behavioral and Brain Sciences</source>
          <volume>3</volume>
          ,
          <fpage>417</fpage>
          -
          <lpage>457</lpage>
          (
          <year>1980</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Koch</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Massimini</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Boly</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tononi</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          :
          <article-title>Neural correlates of consciousness: progress and problems</article-title>
          .
          <source>Nat Rev Neurosci</source>
          <volume>17</volume>
          ,
          <fpage>307</fpage>
          -
          <lpage>321</lpage>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Floridi</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>The Method of Levels of Abstraction</article-title>
          .
          <source>Minds and Machines</source>
          <volume>18</volume>
          ,
          <fpage>303</fpage>
          -
          <lpage>329</lpage>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Gamez</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <source>Are Information or Data Patterns Correlated with Consciousness? Topoi</source>
          <volume>35</volume>
          ,
          <fpage>225</fpage>
          -
          <lpage>239</lpage>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Gamez</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Can we Prove that there are Computational Correlates of Consciousness in the Brain?</article-title>
          <source>Journal of Cognitive Science</source>
          <volume>15</volume>
          ,
          <fpage>149</fpage>
          -
          <lpage>186</lpage>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Chrisley</surname>
          </string-name>
          , R.: Taking Embodiment Seriously:
          <article-title>Nonconceptual Content and Robotics</article-title>
          . In: Ford,
          <string-name>
            <given-names>K.M.</given-names>
            ,
            <surname>Glymour</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            ,
            <surname>Hayes</surname>
          </string-name>
          , P.J. (eds.)
          <article-title>Android Epistemology</article-title>
          . AAAI Press/ The MIT Press, Cambridge and London (
          <year>1995</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Nishimoto</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vu</surname>
            ,
            <given-names>A.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Naselaris</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Benjamini</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gallant</surname>
            ,
            <given-names>J.L.</given-names>
          </string-name>
          :
          <article-title>Reconstructing visual experiences from brain activity evoked by natural movies</article-title>
          .
          <source>Current Biology</source>
          <volume>21</volume>
          ,
          <fpage>1641</fpage>
          -
          <lpage>1646</lpage>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Balduzzi</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tononi</surname>
          </string-name>
          , G.:
          <article-title>Qualia: the geometry of integrated information</article-title>
          .
          <source>PLoS Computational Biology</source>
          <volume>5</volume>
          ,
          <issue>e1000462</issue>
          (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Oizumi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Albantakis</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tononi</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          :
          <article-title>From the phenomenology to the mechanisms of consciousness: Integrated Information Theory 3.0</article-title>
          .
          <source>PLoS Computational Biology</source>
          <volume>10</volume>
          ,
          <issue>e1003588</issue>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>