<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Conceptual Primitive Decomposition for Knowledge Sharing via Natural Language</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jamie C. MACBETH</string-name>
          <email>jmacbeth@fair</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Electrical and Computer Systems Engineering, Fairfield University</institution>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Natural language is an ideal mode of interaction and knowledge sharing between intelligent computer systems and their human users. But a major problem that natural language interaction poses is linguistic variation, or the “paraphrase problem”: there are a variety of ways of referring to the same idea. This is a special problem for intelligent systems in domains such as information retrieval, where a query presented in natural language is matched against an ontology or knowledge base, particularly when its representation uses a vocabulary based in natural language. This paper proposes solutions to these problems in primitive decomposition methods that represent concepts in terms of structures reflecting low-level, embodied human cognition. We argue that this type of representation system engenders richer relations between natural language expressions and knowledge structures, enabling more effective interactive knowledge sharing.</p>
      </abstract>
      <kwd-group>
        <kwd />
        <kwd>Natural language interaction</kwd>
        <kwd>primitive decomposition</kwd>
        <kwd>Conceptual Dependency</kwd>
        <kwd>natural language understanding</kwd>
        <kwd>knowledge sharing</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Natural language promises to be the most effective and convenient method of
transmitting and sharing of knowledge between humans and intelligent computing systems, and
building computing systems that understand natural language as input is an essential part
to having to natural language as a mode of interaction. While the application of machine
learning and deep learning methods to large collections of natural language text have
lead to major successes in natural language processing (NLP), many have noted that the
space of applications that are targeted by these efforts misses out on the broad range of
capabilities that humans exhibit in natural language [8,6].</p>
      <p>Ontologies and knowledge bases will play multiple roles in creating interactive
computing systems that work with natural language. Firstly, to approach applications like
question answering, information retrieval, and narrative understanding, natural language
understanding systems will need to go beyond surface levels of language to access and
apply ontological world knowledge so that they can make inferences and deductions and
draw conclusions similar to those that humans can draw [15,11]. Secondly, ontologies
can be used to represent the knowledge that is shared interactively via the natural
language understanding system.</p>
      <p>But even when knowledge bases and ontological resources are available, a major
challenge of knowledge sharing in the natural language understanding domain is the
“paraphrase problem,” also known in linguistics by the term linguistic variation: the fact
that there are many different ways of saying the same thing [14]. While this term applies
to cross-linguistic variation, or variation between different languages, it also applies to
cases of “free variation” in a single language, where differences in form do not
correspond to any difference in meaning [14, pp. 7–8]. This is a crucial issue for environments
and intelligent systems that share knowledge through interactions in natural language,
because the knowledge about a particular thing may be referred to in many different
ways.</p>
      <p>Some ontologies and knowledge base systems (such as Cyc [7]) may also fall prone
to these problems, because the knowledge about a particular thing is often indexed by
lexical entries, words or phrases, related to that thing. This brief paper proposes research
in a particular class of solutions to paraphrase and linguistic variation problems in natural
language understanding and ontology representations. By formulating a representation
that avoids lexical items and represents knowledge and meaning by decomposing them
into structures that reflect primitive, low-level components of embodied human
perception and cognition, it may enable rich relations between lexical items, varieties of
linguistic expressions, and ontological structures that make for more effective and interactive
natural language understanding systems.</p>
    </sec>
    <sec id="sec-2">
      <title>2. The “Paraphrase Problem” and Linguistic Variation</title>
      <p>
        While a significant literature addresses the challenge of generating paraphrases [9],
detecting paraphrases in the face of linguistic variation serves to be a more challenging
problem, because it requires understanding the language in each paraphrase. To illustrate,
we take the following example problem in story matching and story understanding in
which the system processes stories to make a determination about whether they describe
similar events:
(
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) Bob bought a loaf of bread from the store for $1.
      </p>
      <p>
        (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) The store sold a loaf of bread to Bob for $1.
      </p>
      <p>
        An information retrieval system or a story search and matching system might take (
        <xref ref-type="bibr" rid="ref1">1</xref>
        )
as input, and, given a collection of stories would need to select (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) from the story
collection as a paraphrase of (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ). Alternatively, the problem could be posed as a question
answering problem, where the system is presented with the story (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) as input along with
the question:
      </p>
      <p>
        (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ) Who sold the loaf of bread to Bob?
Determining that (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) describes the same event or set of events as (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) depends crucially
on the system understanding that the same sequence or collection of acts occurred. An
understanding system based naively on analyzing the sentence syntactically and
determining that the verb in (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) is “buy” and the verb in (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) is “sell” might determine that the
sentences did not describe the same events because none of the verbs matched.
      </p>
      <p>A more advanced natural language understanding system might search a knowledge
base or an ontology for propositions stored under the verb “buy” and for relations
between the word “buy” and the word “sell”. If the knowledge base is represented as an
ontology or a semantic network, there may be one or more links or a path consisting of
a series of links in the network in between buying and selling concepts to represent that
they are related.</p>
      <p>Continuing the example, an ontology system such as Text2Onto [2] could be used to
extract ontological concepts and relations from textual and linguistic resources to build
a story understanding system. Text2Onto could be run on a lexical database like
WordNet [3], which is frequently used for text analysis and AI applications. WordNet
organizes an English lexicon primarily into synonym sets or synsets—groupings of words
“that are interchangeable in some context.” These relations include those such as
synonymy/antonymy, hypernymy/hyponymy, meronymy/holonymy and entailment.
WordNet and lexical resources like it should have a synset relation between “buy” and “sell”.</p>
      <p>
        But synset relations alone do not appear to capture the relationships of the actors
and both direct and indirect objects to the events in (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) and (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ). One assumes that a
synonym relation between words implies that one can be replaced with the other in the
correct context. Using the synonym relation to test the story-similarity of (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) and (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) by
replacing “sell” with “buy” in (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) yields:
      </p>
      <p>
        (
        <xref ref-type="bibr" rid="ref4">4</xref>
        ) The store bought a loaf of bread to Bob for $1.
which is obviously not a paraphrase for (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) or (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) and does not lead to the conclusion
that (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) is a paraphrase of (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ).
      </p>
      <p>
        One might think that an antonym relation is more accurate. Since an antonym
relation implies that “buy” is the opposite act of “sell”, testing that (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) is a paraphrase of
(
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) would require a more complex transformation of the sentence, replacing “sell” with
“buy” and switching the actor and the indirect object, and switching the preposition “to”
to “from”. This more complex transformation appears to work in mapping (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) to (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ).
However, a third story:
      </p>
      <p>
        (
        <xref ref-type="bibr" rid="ref5">5</xref>
        ) Bob gave the store $1, and the store gave Bob a loaf of bread.
appears to be a paraphrase of both (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) and (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ), but our synset method will definitely not
work. While it is clear that the act of “giving” is an integral part of buying and selling,
none of the commonly used linguistic synset relations seems adequate in capturing the
relationship. For example, if we try to represent that giving is synonymous with selling, a
transformation of the selling sentence might produce a reasonable-sounding paraphrase
such as
      </p>
      <p>
        (
        <xref ref-type="bibr" rid="ref6">6</xref>
        ) The store gave a loaf of bread to Bob for $1.
      </p>
      <p>
        But it would also produce a number of clearly incorrect paraphrases for matching and
understanding such as these:
(
        <xref ref-type="bibr" rid="ref7">7</xref>
        ) Bob gave a loaf of bread from the store for $1.
(
        <xref ref-type="bibr" rid="ref8">8</xref>
        ) Bob bought the store $1, and the store bought Bob a loaf of bread.
(
        <xref ref-type="bibr" rid="ref9">9</xref>
        ) Bob sold the store $1, and the store sold Bob a loaf of bread.
(
        <xref ref-type="bibr" rid="ref10">10</xref>
        ) Bob bought the store $1, and the store sold Bob a loaf of bread.
(
        <xref ref-type="bibr" rid="ref11">11</xref>
        ) Bob sold the store $1, and the store bought Bob a loaf of bread.
      </p>
      <p>FrameNet [1] is another resource that is frequently used in natural language
understanding systems which, at a glance, may appear to solve this problem. FrameNet links
word senses to each other in frames [4] that are meant to represent the complex
relationships between nouns and verbs in natural language expressions describing a particular
social situation. FrameNet has a Commercial transaction frame which has
semantic and syntactic valence relationships between verbs such as “buy”, “sell”, “charge”,
“spend”, “pay”, and “cost”, and nouns such as “goods” and “money”, as well as
relationships between actors and a Commercial transaction event, such as “buyer” and
“seller”.</p>
      <p>However, when we perform a close examination of the FrameNet frames2 for
Commerce buy, Commerce sell, and Giving, the Commercial transaction
subframes that one might naturally consult in the context of our story matching problem, we
encounter similar issues when we try to generate a relationship between several different
ways of describing the same set of events. The Giving frame in FrameNet describes the
situation using information about the syntactic and semantic relations between words like
Donor, Recipient, and Theme, while the Commerce buy and Commerce sell frames
use a different set of terms: Buyer, Seller, Money, and Goods. Although FrameNet
does define an inheritance relationship between these three frames in which both the
Commerce buy and Commerce sell frames inherit from the Giving frame, no other
annotations in FrameNet exist to indicate, for example, that buying or selling something
involves two acts of giving, or that the donor in the giving act where money was the
theme is the buyer.</p>
      <p>In building automated systems to understand and relate natural language expressions
about similar events through FrameNet frames, we will encounter problems similar to
those we encountered using WordNet synsets. Ultimately, this appears to be because the
content of FrameNet frames is composed of linguistic forms and lexical items which
are subject to the same paraphrasing and linguistic variation that are inherent in natural
language generally.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Solutions in Cognitive Primitive Decomposition</title>
      <p>The preceding examples demonstrate a number of issues with building natural language
understanding systems based on lexical relations and ontologies built from text resources,
given the “paraphrase problem”. In cases where linguistic knowledge bases have
established relations such as synonymy or antonymy, they may also establish rules for
transforming, for example, stories about buying into stories about selling. However, scaling
these systems may require a large number of transformation rules. In the case above,
rules or relations are needed for switching instances of the pair of verbs “buy” and “sell”,
for exchanging the subject with the indirect object. Even more rules are needed for all of
the various permutations involving “give”, “take” and other similar verbs.</p>
      <p>Similarly, while typical ontological relations—such as is-a, instance-of, part-whole,
and equivalence—are very relevant to relations between certain types of objects and
concepts, they do not appear to capture the complex conceptual relationships in
ques2An index of FrameNet frames can be found at https://framenet.icsi.berkeley.edu/fndrupal/frameIndex</p>
      <sec id="sec-3-1">
        <title>STORE ks</title>
      </sec>
      <sec id="sec-3-2">
        <title>3+ ATRANS o</title>
        <p>TJ
r
obj
obj
$1
BOB ks</p>
      </sec>
      <sec id="sec-3-3">
        <title>3+ ATRANS o</title>
        <p>recip
recip
o
o
/ BOB</p>
        <p>STORE
/ STORE</p>
        <p>BOB
“Bob bought a loaf of bread from the store for $1,” “the store sold a loaf of bread to Bob for $1”, and “Bob
gave the store a $1, and the store gave Bob a loaf of bread.” Double arrows represent the relationship between
the animate actor and the ATRANS primitive act, while single arrows marked “obj”, “recip”, indicate the
object and the recipient case of the ATRANS act, respectively. The triple arrow marked “r” represents “result
causation”, indicating that Bob’s act of ATRANSing $1 to the store caused the store to ATRANS a loaf of
bread to Bob.
tion here. Although one could devise unique labels for links between “buy”, “sell”, and
“give”, or place intermediate nodes between them in an ontological network, this
appears to proliferate labels of relations, just as it would transformation rules in a lexical
database.</p>
        <p>
          A potential solution to these problems is to work to represent meaning in a way
that is not only non-linguistic, but reflects imagery, image schemas, or mental models
[12,10,5] composed of the primitive perceptual and cognitive experiences that people
use when processing and understanding language. Figure 1 shows a conceptual diagram
meant to represent the concept behind the three paraphrases (
          <xref ref-type="bibr" rid="ref1">1</xref>
          ), (
          <xref ref-type="bibr" rid="ref2">2</xref>
          ), and (
          <xref ref-type="bibr" rid="ref5">5</xref>
          )
simultaneously. The conceptual diagram is composed in a system called Conceptual Dependency
[13], which attempts to decompose the concept behind the paraphrases into a complex
combination of language-free conceptual primitives; in this case using a conceptual
primitive called ATRANS, which represents an event that changes an abstract relationship
between human actors and an inanimate object. While the top ATRANS primitive act
in the diagram represents a conceptualization of the store transferring possession of the
bread to John, the bottom ATRANS primitive act represents bob transferring possession
of the $1 to the store.
        </p>
        <p>A story understanding system tasked with relating the given paraphrases can attempt
to decompose each the different sentences into this “conceptual base” form, and perform
a comparison of the conceptual diagrams to determine that they describe the same set
of events. If the representation system uses a small number of conceptual primitives and
connectives, but combines them in complex ways to correspond to the variation of natural
language expressions, the matching occurs through relatively simple processes of graph
isomorphism and structure mapping.</p>
        <p>If the problem with synset relations and standard ontology relations is that they will
proliferate a complex array of named relations between words or between concepts, we
foresee that methods of decomposing words into complex combinations of primitives
creates a substrate to capture the complex relations between words and other words,
between words and concepts, and between concepts and other concepts. Ultimately,
decomposing concepts in this way allows the decomposition into cognitive primitives itself
to define the system of relations. This appears to be a more realistic representation of
human cognition behind the relations between concepts than a simple system of labeled
edges.</p>
        <p>We propose further research in supplementing or enhancing available ontologies,
knowledge bases, and lexical databases with resources that present knowledge in a form
that is decomposed into structured combinations of conceptual primitives. We argue that
this is essential to achieving richer and more in-depth forms of computer-based natural
language understanding. This, in turn, may enable greater knowledge sharing between
people and all systems that represent knowledge in the form of natural language.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Collin</surname>
            <given-names>F Baker</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Charles J Fillmore</surname>
            ,
            <given-names>and John B Lowe.</given-names>
          </string-name>
          <article-title>The berkeley FrameNet project</article-title>
          .
          <source>In Proceedings of the 36th Annual Meeting of the Association for Computational Linguistics and 17th International Conference on Computational Linguistics-Volume</source>
          <volume>1</volume>
          , pages
          <fpage>86</fpage>
          -
          <lpage>90</lpage>
          . Association for Computational Linguistics,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Philipp</given-names>
            <surname>Cimiano</surname>
          </string-name>
          and
          <article-title>Johanna Vo¨lker. A framework for ontology learning and data-driven change discovery</article-title>
          .
          <source>In Proceedings of the 10th International Conference on Applications of Natural Language to Information Systems (NLDB)</source>
          , volume
          <volume>3513</volume>
          of Lecture Notes in Computer Science, pages
          <fpage>227</fpage>
          -
          <lpage>238</lpage>
          . Springer,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Christiane</given-names>
            <surname>Fellbaum. WordNet</surname>
          </string-name>
          :
          <article-title>An Electronic Lexical Database</article-title>
          . MIT Press,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Charles</surname>
            <given-names>J</given-names>
          </string-name>
          <string-name>
            <surname>Fillmore and Beryl T Atkins</surname>
          </string-name>
          .
          <article-title>Toward a frame-based lexicon: The semantics of RISK and its neighbors</article-title>
          .
          <source>In Frames, Fields, and Contrasts: New Essays in Semantics and Lexical Organization</source>
          , pages
          <fpage>75</fpage>
          -
          <lpage>102</lpage>
          . Lawrence Erlbaum Associates, Hillsdale, NJ,
          <year>1992</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Philip</given-names>
            <surname>Nicholas</surname>
          </string-name>
          Johnson-Laird.
          <article-title>Mental models: Towards a Cognitive Science of Language, Inference, and Consciousness</article-title>
          . Harvard University Press,
          <year>1983</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Pat</given-names>
            <surname>Langley</surname>
          </string-name>
          .
          <article-title>Intelligent behavior in humans and machines</article-title>
          .
          <source>Advances in Cognitive Systems</source>
          ,
          <volume>2</volume>
          :
          <fpage>3</fpage>
          -
          <lpage>12</lpage>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>D. B.</given-names>
            <surname>Lenat</surname>
          </string-name>
          and
          <string-name>
            <given-names>R. V.</given-names>
            <surname>Guha</surname>
          </string-name>
          .
          <article-title>Building Large Knowledge-Based Systems</article-title>
          . Addison-Wesley, Reading, MA,
          <year>1990</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Hector</surname>
            <given-names>J</given-names>
          </string-name>
          <string-name>
            <surname>Levesque</surname>
          </string-name>
          .
          <article-title>On our best behaviour</article-title>
          .
          <source>Artificial Intelligence</source>
          ,
          <volume>212</volume>
          :
          <fpage>27</fpage>
          -
          <lpage>35</lpage>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Nitin</given-names>
            <surname>Madnani</surname>
          </string-name>
          and
          <string-name>
            <given-names>Bonnie J.</given-names>
            <surname>Dorr</surname>
          </string-name>
          .
          <article-title>Generating phrasal and sentential paraphrases: A survey of datadriven methods</article-title>
          .
          <source>Comput. Linguist.</source>
          ,
          <volume>36</volume>
          (
          <issue>3</issue>
          ):
          <fpage>341</fpage>
          -
          <lpage>387</lpage>
          ,
          <year>September 2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Jean</surname>
            <given-names>M</given-names>
          </string-name>
          <string-name>
            <surname>Mandler and</surname>
          </string-name>
          <article-title>Cristo´bal Paga´n Ca´novas. On defining image schemas</article-title>
          .
          <source>Language and Cognition</source>
          ,
          <volume>6</volume>
          (
          <issue>4</issue>
          ):
          <fpage>510</fpage>
          -
          <lpage>532</lpage>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Vincent</given-names>
            <surname>Ng</surname>
          </string-name>
          .
          <article-title>Machine learning for entity coreference resolution: A retrospective look at two decades of research</article-title>
          .
          <source>In Proceedings of the Thirty-First AAAI Conference on Artificial Intelligence</source>
          , pages
          <fpage>4877</fpage>
          -
          <lpage>4884</lpage>
          . AAAI,
          <year>2017</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M.</given-names>
            <surname>Sadoski</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Paivio</surname>
          </string-name>
          .
          <article-title>Imagery and Text: A Dual Coding Theory of Reading and Writing</article-title>
          . Taylor &amp; Francis, New York,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Roger</surname>
            <given-names>C</given-names>
          </string-name>
          <string-name>
            <surname>Schank. Conceptual Information Processing. Elsevier</surname>
          </string-name>
          , New York, NY,
          <year>1975</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>James</surname>
            <given-names>A Walker.</given-names>
          </string-name>
          <article-title>Variation in Linguistic Systems</article-title>
          . Routledge, New York, NY,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Patrick</surname>
            <given-names>Henry</given-names>
          </string-name>
          <string-name>
            <surname>Winston</surname>
          </string-name>
          .
          <article-title>The genesis story understanding and story telling system: A 21st century step toward artificial intelligence</article-title>
          .
          <source>Technical report</source>
          , Center for Brains,
          <source>Minds and Machines (CBMM)</source>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>