<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>The Roles of Symbols in Neural-based AI: They are Not What You Think!</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Daniel L. Silver</string-name>
          <email>danny.silver@acadiau.ca</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tom M. Mitchell</string-name>
          <email>tom.mitchell@cs.cmu.edu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Jodrey School of Computer Science, Acadia University</institution>
          ,
          <addr-line>Wolfville, NS</addr-line>
          ,
          <country country="CA">Canada</country>
          <addr-line>B4P2R6</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Machine Learning Department, Carnegie Mellon University</institution>
          ,
          <addr-line>Pittsburg, PA, USA, 15213</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>NeSy 2023, 17th Intern. Workshop on Neural-Symbolic Learning and Reasoning, Certosa di Pontignano</institution>
          ,
          <addr-line>Siena</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <abstract>
        <p>We present a novel neuro-symbolic hypothesis and an architecture for intelligent agents that combines subsymbolic representations for symbols and concepts for learning and reasoning. We argue that symbols will remain critical to the future of intelligent systems NOT because they are the fundamental building blocks of thought, but because they characterize the subsymbolic processes that constitute thought. In [1] we begin by defining terminology for discussing the neural encoding of symbols and concepts, and describing the key questions we seek to answer about neuro-symbolic systems. We then present relevant research results from neuroscience, behavioral (cognitive) science, and artificial intelligence, that yield evidence about the combination of symbolic and subsymbolic processing in humans and current artificial neural networks. Guided by this evidence, we present a novel neuro-symbolic hypothesis and an associated architecture meant to provide a plausible answer to the question of how humans might implement neuro-symbolic reasoning, and how future intelligent agents might be designed to do so as well.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Definitions:
string “peach”. In the paper, we present results from brain imaging studies that reveal some of
the properties and timing of these symrep and conrep patterns of neural activity.</p>
      <p>
        A Neuro-Symbolic Hypothesis: Symbols are critical to intelligence NOT because they are
the building blocks of thought, but because they are characterizations of thought that (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) allow us
to explain our subsymbolic thinking to ourselves and others and (2) act as constraints on inference
and learning about the world. Symbols explain our thinking and aid our thinking, but are not the
foundation of our thinking.
      </p>
      <p>
        We propose a neuro-symbolic architecture for an intelligent agent such as a human. It borrows
from the basic diagram of an Intelligent Agent as well as work in neuroscience, cognitive
psychology (specifically the ideas of Thinking Fast and Slow by D. Kahneman), and deep neural
networks. It is also inspired by recent neuro-symbolic literature (particularly by L. Lamb and A.
Garcez). The four major components of the architecture are as follows: Sensory and Motor
Subsystems that receive raw external percepts (images, sounds, touch) from the real world
and provide the appropriate percept signals to the higher order System 1 and 2 components.
A System 1 attractor network that, given a percept signal, a symrep vector which contains
recent context and attention information, and a goal-driven attention vector, learns to relax
into a conrep activation state. This conrep will represent one or more of the previously learned
concepts, each at some appropriate level of abstraction. It is semantically organized and largely
grounded in sensory/motor representations with the properties: (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) two concepts with similar
meanings have similar representations, and (2) operations over pairs of conrep vectors (e.g.
superposition) can be performed using mapping functions. A System 2 attractor network that,
given a percept signal, a conrep vector which contains recent context and attention information,
and a goal-driven attention vector, learns to relax into a desired symrep activation state. The
symrep will represent one or more of the previously learned symbols or some proto-symbol
(discussed in the paper). System 2 is organized so that (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) symbols with similar appearance have
proximal symreps, and (2) symrep operations and composition can be done with vector mapping
functions (e.g. 7 × 9 =&gt; 63, ”fish” + ”ing” = ”fishing”). A Performance Goal Subsystem that
influences the formation of subsequent conrep and symrep given the agent’s recent perceptual
input and internal state. Note that performance goals (e.g. to eat) can be closely associated
with internal senses (e.g. hunger). These goals, context and attention vectors driven by recent
conrep, have influence on the training and relaxation of the Systems 1 and 2 attractor networks.
To some extent, agents of this architecture see what they want to see from the percepts because
it helps them make sense of their world.
      </p>
      <p>
        Conclusion: We conjecture that internal agent “self-communication” using the symrep
meant for agent-to-agent communication, has become key to human intelligence because: (
        <xref ref-type="bibr" rid="ref1">1</xref>
        )
it provides a second, more abstract level of representation and reasoning which can occur in
parallel with subsymbolic reasoning, and (2) it places an additional constraint on learning where
prior learning act as an inductive bias for learning new symbols and concepts. Shared symbols
allow us to explain and justify, internally as well as externally, our decisions and actions. And
what we learn is shaped and constrained by the “lexicon” of what we recognize as symbols.
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>D. L.</given-names>
            <surname>Silver</surname>
          </string-name>
          , T. M. Mitchell,
          <article-title>The roles of symbols in neural-based AI: They are not what you think!</article-title>
          , In: P. Hitzler,
          <string-name>
            <given-names>M. K.</given-names>
            <surname>Sarker</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Eberhart (Eds.),
          <source>Compendium of Neuro-Symbolic Artificial Intelligence</source>
          , IOS Press, Amsterdam,
          <year>2023</year>
          . See: arxiv.org/abs/2304.13626.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>