<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Automating the Encoding for the LISA Model of Analogy from Raw Text</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>SEAN WILNER</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>University of Illinois Urbana-Champaign JE HUMMEL</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>University of Illinois Urbana-Champaign</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>General Terms: Analogy</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Additional Key Words and Phrases: Natural Language Processing</institution>
          ,
          <addr-line>Computational Linguistics, Cognitive Psychology, Analogy</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>is material is based on research sponsored by the Air Force Research Laboratory, under agreement number FA9550-12-1-0003. e U.S. Government is authorized to reproduce and distribute reprints for Governmental purposes notwithstanding any copyright notation thereon. Author's addresses: S. Wilner, Illinois Informatics Institute, University of Illinois UrbanaChampaign; JE Hummel, Department of Psychology, University of Illinois UrbanaChaimpaign</institution>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2017</year>
      </pub-date>
      <volume>1</volume>
      <issue>1</issue>
      <fpage>115</fpage>
      <lpage>122</lpage>
      <abstract>
        <p>Analogy is an integral part of human cognition. Consequently researchers would like to produce computational models of analogy to test implementations of theoretical claims. Such computational models have relied almost exclusively upon hand-coded representations, making the resulting models too dependent upon the modeler's choices, and thus di cult to interpret in isolation of the modeler. In an e ort to combat this dependency, we present here a means of automatic encoding for the LISA model of analogy and inference.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 INTRODUCTION</title>
      <p>
        Analogy, and more generally, relational reasoning are core cognitive capacities [
        <xref ref-type="bibr" rid="ref10 ref6 ref8">6, 8, 10</xref>
        ] that, along with language,
seem to separate human cognition from the cognitive abilities of all the other great apes [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ]. Many of the
computational models of analogy arising from machine learning revolve around solving semantic, or 4-way,
analogies of the form “king:queen::brother:?” where the model lls in “sister” [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. ese models fail to account
for the more substantial form of analogy – narrative analogy – wherein agents and actions must all be analogically
mapped to one another. Numerous models have been developed to account for our ability to make such analogies.
Among the most in uential have been Falkenhainer et al.’s [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] Structure Mapping Engine (SME), Holyoak and
agard’s [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] Analogical Constraint Mapping Engine (ACME) and Hummel and Holyoak’s [
        <xref ref-type="bibr" rid="ref11 ref12">11, 12</xref>
        ] Learning and
Inference with Schemas and Analogies (LISA) (see Gentner &amp; Forbus, 2011[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], for a review).
      </p>
      <p>
        Although these and other models account for many aspects of human relational thought, they share a common
limitation: All of them rely on hand-coded knowledge representations, placing a great deal of control in the
modeler’s hands. As such, it is o en unclear how many of the successes of these models re ect the models’
fundamental tenets and how many re ect arbitrary representational and modeling choices made by the modeler
at the time of creating the simulations (see, e.g., Lu et al. 2008 [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]). Since such models are only as good as their
modeler, the support or challenge to any theoretical claims they make must be interpreted carefully.
      </p>
      <p>
        Ideally, these models would produce their representational inputs directly from natural language, removing
their dependence upon hand-coded inputs. ere are some models that approach this goal for speci c narrow
domains, such as intuitive physics (Friedman SE &amp; Forbus KD, 2009 [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]) and moral decision making (Dehghani M
et al., 2008 [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]). However, the task of generalizing open domain natural language for analogy remains elusive. To
address this problem, we describe recent e orts applying several Natural Language Processing (NLP) techniques
c 2016 ACM. /2016/1-ART1 $
DOI:
1:2 •
to the automatic generation of knowledge representations for LISA. When combined, these techniques allow us
to transform English sentences into semantically-rich propositional representations usable to the LISA model of
analogy (i.e., representations in “LISAese”).
1.1
      </p>
    </sec>
    <sec id="sec-2">
      <title>Representation</title>
      <p>LISAese is a collection of semantic encodings and structural representations which, together, allow LISA to
produce semantically informed, structurally consistent analogies. LISA employs semantic representations for
both objects and predicates and encodes structure in a predicated logical form. Semantics for an object in LISAese
consist of a distributed list of abstract “semantic units” and corresponding weights. For instance, if we were to
hand-code the representation for a taxi driver, “Bob” and a bank teller “Jill”, we may end up with the following
semantics and corresponding weights:</p>
      <p>Bob: human=1.0, adult=0.7, male=1.0, employed=0.8, taxi driver=1.0
Jill: human=1.0, adult=0.7, female=1.0, employed=0.8, bank teller=1.0</p>
      <p>Bob’s semantics could then overlap with the semantics for our bank teller, “Jill”, who is also human, adult, and
employed, but would di er on male and taxi driver. LISA treats these “semantic units” as universal across all
representations, that is, if two “semantic units” have the same name, they are the same unit. Aside from this
universality, the names are arbitrary and the choice to give them relevant names with regard to the objects they
are a ached to only serves to facilitate human readability.</p>
      <p>Similarly, predicate semantics consist of “semantic units” and weights; however, predicate semantics are de ned
at the level of their relational roles rather than the predicate itself. For example, the two-place predicate love is
de ned by two lists of semantics. e rst list represents the agent role of love (the lover); the second represents
the patient role of love (the beloved). is could be represented as follows:</p>
      <p>love:
[ lover=1.0, emotion=1.0, strong emotion=0.8, passionate=0.9 ]
[ beloved=1.0, receives emotion=1.0, receives strong emotion=0.8, receives passion=0.9 ]
With de ned semantics on objects and predicates, LISA is able to ground its structural representation, given in
a predicated logical format, to represent “Bob loves Jill”: love(Bob, Jill). is logical representation is robust to
recursive references such as “Jill knows that Bob loves her.” which yields: know-(Jill, love(Bob, Jill)). Indeed, this
is su cient to represent the necessary semantic and structural information LISA needs to make analogies and
inferences.
1.2</p>
    </sec>
    <sec id="sec-3">
      <title>Task</title>
      <p>Our task in this paper is to outline a technique to produce semantic and structural representations automatically
from text. Subsequently, we will show that LISA can use the resulting LISAese to successfully nd analogies and
inferences. More speci cally, we will present a method which achieves the following:
(1) Generates Semantics
• Object semantics that account for modi cation by adjectives, noun-phrases, and copular verb
a achments
• Predicate semantics with role-by-role representation
(2) Generate Structure</p>
    </sec>
    <sec id="sec-4">
      <title>Automating the Encoding for the LISA Model of Analogy from Raw Text • 1:3</title>
      <p>• Produces logic form
• Handles recursive references
• Associates multiple references to a single entity
By using NLP techniques to automatically pre-process and encode text into LISAese, the LISA model will no
longer need to rely on hand-coded knowledge representations. Freeing it of this requirement will clarify how
LISA’s successes as a model re ect its fundamental tenets rather than choices made by the modeler.
2</p>
    </sec>
    <sec id="sec-5">
      <title>GENERATE SEMANTICS</title>
      <p>Ultimately, we need to generate semantic representations for both objects and predicates, but the requirements
for automated generation of the two di er slightly. is is because, as previously noted, in LISAese semantics are
assigned not to the predicates themselves, but rather to each predicate role independently.
2.1</p>
    </sec>
    <sec id="sec-6">
      <title>Object Semantics</title>
      <p>
        Objects may be represented in a fairly straightforward manner using word embedding. Word embedding is a
means of mapping words into a semantic space (a vector space) such that words which are ‘similar’ are closer in
the resulting space than words which are ‘dissimilar’. ere are many means of producing word embeddings.
One of the most commonly used semantic embedding tools is Google’s Word2Vec [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] which produces vectors
based o carefully cra ed word co-occurrence statistics. Using these vectors, we can produce embeddings for
individual words such as “cat”, “January”, or “happy”. However, this means of embedding can break down upon
encountering proper nouns such as “Microso ” or “Bob”. In order to properly encode proper nouns, we employ an
NLP technique called Named Entity Recognition (NER) which identi es what kind of entity the proper nouns are:
“Microso ”!ORGANIZATION and “Bob”!PERSON. Using these NER tags, we can encode the proper nouns by
the semantic embedding of their associated tag (for added speci city, where known, we also incorporate the
gender of each proper noun).
      </p>
      <p>In order to accommodate various modi ers, we additively combine the semantic vectors corresponding to each
object. ese vectors consist of the embeddings of the words contained within or associated with an entity, for
example, the words in a noun-phrase, an NER label, adjective modi ers, or copular modi ers. Copular verbs
are stative verbs such as is/are/was/were, seem, or appear. For example, “smelly Sam” and “Sam is smelly” both
result in the incorporation of the semantics for “smelly” in the representation of Sam.</p>
      <p>Using the resulting vectors, we can produce LISAese by creating an arbitrary semantic label, call it “SemUnit i”,
to every ith index of the vector where the weight on the generated semantic unit is the value at the ith index of
the vector. For example, if our word embedding was 4 dimensions (instead of 300), we could have the following
embedding for “happy”:
happy! [0.14, 0.0024, 0.81, -0.37]</p>
      <sec id="sec-6-1">
        <title>Which would then yield a LISAese representation of “happy”:</title>
        <p>happy! SemUnit 0=0.14, SemUnit 1=0.0024, SemUnit 2=0.81, SemUnit 3=-0.37
us allowing us to produce object semantics in LISAese from the Word2Vec word-embeddings.
1:4 •
Predicate semantics in their most general form are more di cult to automatically encode than objects. Direct
role-by-role representation via word-embedding is certainly non-obvious, although we will explore methods to
do so in the future work section. Here, we will simplify our semantic representation to the easiest usable form
wherein we assign semantics to each role corresponding to word embeddings for the predicate. For example, in
our 4-dimensional embedding from before, we can imagine the following vector for “love”:
”love” ! [0.23, 0.01, 0.76, -0.45]</p>
      </sec>
      <sec id="sec-6-2">
        <title>Copying the embedding onto the agent and patient roles of love gives us:</title>
        <p>love:
[SemUnit 0agent=0.23, SemUnit 1agent=0.01, SemUnit 2agent=0.76, SemUnit 3agent=-0.45]
[SemUnit 0patient=0.23, SemUnit 1patient=0.01, SemUnit 2patient=0.76, SemUnit 3patient=-0.45]
While this representation meets all of our requirements, it does have a glaring weakness. ere will never be
any semantic overlap between two di erent predicate roles (e.g. agent and patient). e resulting LISAese will
make LISA preferentially map agents and patients onto other agents and patients respectively. Again, we will
explore alternatives to this encoding which do not su er from the same weakness in the future works section
below.</p>
        <p>Using the above semantic representations, we meet all the requirements for LISAese’s semantic encoding and
are ready to explore how to structurally represent text in LISAese.
3</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>GENERATE STRUCTURE</title>
      <p>
        In order to encode language in the kind of predicated logic LISAese requires, we need to identify predicate roles
and the objects which ll them. In the simplest case, we may be presented with text such as “Bob loves Jill”.
In which there is one predicate with two roles, agent and patient, lled by “Bob” and “Jill” respectively. As
mentioned earlier, this structure becomes more di cult when predicates ll the roles of other predicates as in
“Jill knows Bob loves her”. In this later case, the structure we hope to capture is that “Jill” is the agent and love
is the patient of know. To discover these a achments, we may employ an NLP technique called Semantic Role
Labeling (SRL). SRL identi es predicates and their roles as well as portions of the text which ll those roles. Using
SRL, we can label “Jill” as the agent and “Bob loves her” as the patient of know. Similarly, “Bob” is identi ed
as the agent, and “her” the patient of love. For our work, we use the SRL engine laid out in Punyakanok, Roth,
and Yih [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ] which uses joint inference over several SRL systems and Integer Linear Programming for global
constraint satisfaction to optimize its labeling. A er using our SRL engine, we have the following structural
encoding: know(Jill, love(Bob, her)). is is nearly enough to properly structure the input for LISA since we are
able to encode the predicates in a LISA’s logical form and are able to handle recursive references to predicates
as arguments to other predicates. e last remaining piece is to properly associate “Jill” with “her”. In order
to achieve this, we use another NLP technique called ‘co-reference resolution’ [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]. Co-reference resolution
associates multiple references to the same entity throughout a text. Using this technique, we are able to associate
“her” with “Jill”, yielding our nal representation: know(Jill, love(Bob, Jill)).
      </p>
      <p>e automated structural encoding, along with the semantic encoding, is su cient to produce functional
LISAese without the involvement of any hand-coding and li le modeler input. However, this automatic encoding
is useless if LISA is not able to function on the resulting LISAese. In order to test this, we explore three benchmark
simulations LISA has successfully completed in the past: Latin to Logic, Love Triangle, and Driving a Vehicle.</p>
    </sec>
    <sec id="sec-8">
      <title>Automating the Encoding for the LISA Model of Analogy from Raw Text • 1:5</title>
      <p>4</p>
    </sec>
    <sec id="sec-9">
      <title>SIMULATIONS</title>
      <p>To test our automated representations, we will compare LISA’s discovered mappings and inferences on them
against the hand-coded LISAese. Since our automatic encoding produces several hundred semantics for each
object and role, we have omi ed them from this paper. However, since there were some di erences in some of
the structural representations between the hand-coded and automatic representations, we have provided both for
each benchmark below.</p>
      <sec id="sec-9-1">
        <title>For “Latin to Logic”[14], the benchmark representation is shown below:</title>
      </sec>
      <sec id="sec-9-2">
        <title>Analog 1: is incomplete(Second Order Logic)</title>
      </sec>
      <sec id="sec-9-3">
        <title>Analog 2: is self referential(Natural Language) has unprovable statement(Natural Language) is incomplete(Natural Language)</title>
        <p>LISA can use this produced representation to nd the analogy between Second Order Logic and Natural
Language and infer that Second Order Logic has unprovable statements and that it is incomplete.</p>
        <p>In order to test our automatic representation, we need to identify English sentences which convey a similar
meaning. Running LISA over the representations generated by these sentences, we can check if it produces
the correct mappings and inferences. For that purpose, we chose the following stories: “Second Order Logic is
incomplete.” and “Latin is able to self reference. Latin proves and disproves a statement. Latin is incomplete.”
Using the automatic representational system described here, we produced the following representation:</p>
      </sec>
      <sec id="sec-9-4">
        <title>Analog 1: be(Second Order Logic, incomplete)</title>
      </sec>
      <sec id="sec-9-5">
        <title>Analog 2:</title>
        <p>be(Latin, able to self reference)
prove(Latin, a statement)
disprove(Latin, a statement)
be(Latin, incomplete)</p>
        <p>Running LISA on this representation does in fact produce a mapping between Second Order Logic and Latin.
LISA makes both of the desired inferences, namely that Second Order Logic is self referential and that it can
prove and disprove a statement. Worth noting is that the structure generated by our automated system is quite
di erent from that of the original representation. However, the correct mapping and inference was still produced.</p>
        <p>
          For “Love Triangle” [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] the hand-coded benchmark structure and the automatic structural encoding were
identical, and it is shown below:
        </p>
      </sec>
      <sec id="sec-9-6">
        <title>Analog 1:</title>
        <p>love(Sally, Jack)
1:6 •</p>
        <p>“Latin to Logic” is the easiest of our three benchmarks since the structure is simple and semantic similarity
alone is enough inform mapping. In “Love Triangle”, while the structure remains simplistic, the object semantics
are inversely correlated with the ideal mapping. at ideal mapping is from “Sally”!“John”, “Jack”!“Mary”, and
“Megan”!“Bob”, which is counter-indicated by semantic similarity on the subset of semantics corresponding to
gender. In the hand-coded version, LISA nds this ideal mapping and produces the correct inference, namely
hates(John, Bob) in Analog 2. A er running LISA on the automatically produced LISAese, the same mapping and
inference was found.</p>
        <p>
          Lastly, we will look at the most complicated benchmark, “Driving a Vehicle” [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ]. Its hand coded structure is
shown below:
        </p>
      </sec>
      <sec id="sec-9-7">
        <title>Analog 1: has(Joan, car) want(Joan, goto(Joan, airport))</title>
      </sec>
      <sec id="sec-9-8">
        <title>Analog 2:</title>
        <p>has(Sam, car)
want(Sam, goto(Sam, beach))
drive(Sam, car, beach)
e automatic encoding produced a very similar structure:
Analog 1:
have(Joan, a car)
want(Joan, go(Joan, the airport))</p>
      </sec>
      <sec id="sec-9-9">
        <title>Analog 2: have(Sam, a jeep) want(Sam, go(Sam, to the beach)) drive(Sam, a jeep, to the beach)</title>
        <p>To produce this, the NLP pre-processor was given “Joan has a car. She wants to go to the airport.” for Analog 1
and “Sam has a jeep. He wants to go the beach. He drives the jeep to the beach.” for Analog 2.</p>
        <p>On both the hand-coded and auto-coded LISAese, LISA found the correct mapping that “Joan” is like “Sam”
and the “jeep” is like the “car”. LISA also managed to produce the correct inference that Joan drives her car to the
airport.</p>
      </sec>
    </sec>
    <sec id="sec-10">
      <title>Automating the Encoding for the LISA Model of Analogy from Raw Text • 1:7</title>
      <p>5</p>
    </sec>
    <sec id="sec-11">
      <title>CONCLUSION</title>
      <p>Using NLP techniques to pre-process and encode text into LISAese has proven to be a fruitful approach to
incorporating the encoding process into LISA as a model. With the incorporation of this linguistic front-end, LISA
is now the rst computational model of analogy capable of encoding simple open-domain text and performing
analogical mapping and inference over it. More importantly, the LISA model no longer needs to rely on hand-coded
knowledge representations. With this, it is now clearer how LISA’s successes as a model re ect its fundamental
tenets rather than the representational and modeling choices made by the modeler.
6</p>
    </sec>
    <sec id="sec-12">
      <title>FUTURE WORK</title>
      <p>
        We hope to improve both the semantic and structural encoding in the future with be er semantic libraries and
relational capture. For object semantics, we can use sparse over-complete vectors [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] for the word embeddings,
since it has previously been shown that LISA performs be er over sparse encodings. For predicate semantics, we
can explore role speci c semantics so that predicate role semantics can overlap between labeled roles (some agents
overlap with some patients). Towards that end, we intend to explore several di erent tracks for embedding. We
will look at PropBank [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] labels for verb speci c roles and use the embeddings of the words in their labels (agent
of love is “lover”, patient of love is “loved”). We will also look at embedding the roles directly, either by forming
role co-occurence chains and using Word2Vec on those chains directly, or by replacing the spans of text by their
SRL label and running Word2Vec over the replaced text. However, regardless of what embedding techniques
we use, the representation technique as we’ve described it falls pray to a common word embedding weakness –
inability to di erentiate phrasal cconstructs. at is, “the cat on the blanket” vs “the blanket on the cat” have the
same semantic representation. is pitfall can be avoided by capturing only the head of noun phrase (with
adjectives) and instead predicating any prepositional relationships contained within the phrase. We will also explore
additional ways to extract structure from language in the form of explicitly stated intra-sentential cause and
e ect: “Bob loves Susan because she is kind.” We will store the extracted information in “Cause-E ect” groups [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
      </p>
      <p>
        In addition to improvements of the encoding, we hope to expand our use of LISA. With the availability of
computational resources as well as abundant text, auto-encoding LISAese allows for an explosion of analogical
modeling. As research progresses, we hope to use auto-coded LISA’s mappings and inferences as learning features
on a range of traditional NLP tasks. Success in these domains would demonstrate the importance of analogy
as a core cognitive capacity and its salience to a host of higher level computational tasks, for example schema
generation [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ][
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], the Story Cloze Task [
        <xref ref-type="bibr" rid="ref20 ref21">20, 21</xref>
        ], story classi cation or even provide some insight into the
Winograd Schema challenge [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ].
      </p>
    </sec>
    <sec id="sec-13">
      <title>ACKNOWLEDGMENTS</title>
      <p>is material is based on research sponsored by the Air Force Research Laboratory, under agreement number
FA9550-12-1-0003. e U.S. Government is authorized to reproduce and distribute reprints for Governmental
purposes notwithstanding any copyright notation thereon.</p>
      <p>e authors would also like to thank Hani Awni and Megan Emigh for their input throughout the process.
1:8 •</p>
    </sec>
    <sec id="sec-14">
      <title>Wilner &amp; Hummel</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Niranjan</given-names>
            <surname>Balasubramanian</surname>
          </string-name>
          , Stephen Soderland, Oren Etzioni Mausam,
          <string-name>
            <given-names>and Oren</given-names>
            <surname>Etzioni</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Generating Coherent Event Schemas at Scale.</article-title>
          .
          <source>In EMNLP</source>
          .
          <volume>1721</volume>
          -
          <fpage>1731</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Tomai</given-names>
            <surname>E. Forbus K. Klenk M. Dehghani</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <year>2008</year>
          .
          <article-title>An Integrated Reasoning Approach to Moral Decision-Making. AAAI Conference on Arti cial Intelligence (AAAI) Twenty- ird (</article-title>
          <year>2008</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Brian</given-names>
            <surname>Falkenhainer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Kenneth D</given-names>
            <surname>Forbus</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Dedre</given-names>
            <surname>Gentner</surname>
          </string-name>
          .
          <year>1989</year>
          .
          <article-title>e structure-mapping engine: Algorithm and examples</article-title>
          .
          <source>Arti cial intelligence 41</source>
          ,
          <issue>1</issue>
          (
          <year>1989</year>
          ),
          <fpage>1</fpage>
          -
          <lpage>63</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Manaal</given-names>
            <surname>Faruqui</surname>
          </string-name>
          , Yulia Tsvetkov, Dani Yogatama, Chris Dyer, and
          <string-name>
            <surname>Noah</surname>
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Smith</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Sparse Overcomplete Word Vector Representations</article-title>
          .
          <source>CoRR abs/1506</source>
          .
          <year>02004</year>
          (
          <year>2015</year>
          ). h p://arxiv.org/abs/1506.02004
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Sco</surname>
            <given-names>E</given-names>
          </string-name>
          <string-name>
            <surname>Friedman</surname>
          </string-name>
          ,
          <string-name>
            <surname>Kenneth D Forbus</surname>
            , and
            <given-names>Jason</given-names>
          </string-name>
          <string-name>
            <surname>Taylor</surname>
          </string-name>
          .
          <article-title>Learning and Reasoning with alitative Models of Physical Behavior. ( ).</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Dedre</given-names>
            <surname>Gentner</surname>
          </string-name>
          .
          <year>1983</year>
          .
          <article-title>Structure-mapping: A theoretical framework for analogy</article-title>
          .
          <source>Cognitive science 7</source>
          ,
          <issue>2</issue>
          (
          <year>1983</year>
          ),
          <fpage>155</fpage>
          -
          <lpage>170</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Dedre</given-names>
            <surname>Gentner</surname>
          </string-name>
          and
          <string-name>
            <given-names>Kenneth D</given-names>
            <surname>Forbus</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>Computational models of analogy</article-title>
          .
          <source>Wiley interdisciplinary reviews: cognitive science 2</source>
          ,
          <issue>3</issue>
          (
          <year>2011</year>
          ),
          <fpage>266</fpage>
          -
          <lpage>276</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Mary</surname>
            <given-names>L</given-names>
          </string-name>
          <string-name>
            <surname>Gick and Keith J Holyoak</surname>
          </string-name>
          .
          <year>1983</year>
          .
          <article-title>Schema induction and analogical transfer</article-title>
          .
          <source>Cognitive psychology 15</source>
          ,
          <issue>1</issue>
          (
          <year>1983</year>
          ),
          <fpage>1</fpage>
          -
          <lpage>38</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Keith</surname>
            <given-names>J</given-names>
          </string-name>
          <string-name>
            <surname>Holyoak</surname>
          </string-name>
          and Paul agard.
          <year>1989</year>
          .
          <article-title>Analogical mapping by constraint satisfaction</article-title>
          .
          <source>Cognitive science 13</source>
          ,
          <issue>3</issue>
          (
          <year>1989</year>
          ),
          <fpage>295</fpage>
          -
          <lpage>355</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Keith</given-names>
            <surname>James Holyoak</surname>
          </string-name>
          and Paul agard.
          <year>1995</year>
          .
          <article-title>Mental leaps</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>John</surname>
            <given-names>E</given-names>
          </string-name>
          <string-name>
            <surname>Hummel and Keith J Holyoak</surname>
          </string-name>
          .
          <year>1997</year>
          .
          <article-title>Distributed representations of structure: a theory of analogical access and mapping</article-title>
          .
          <source>Psychological Review</source>
          <volume>104</volume>
          ,
          <issue>3</issue>
          (
          <year>1997</year>
          ),
          <fpage>427</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>John</surname>
            <given-names>E</given-names>
          </string-name>
          <string-name>
            <surname>Hummel and Keith J Holyoak</surname>
          </string-name>
          .
          <year>2003</year>
          .
          <article-title>A symbolic-connectionist theory of relational inference and generalization</article-title>
          .
          <source>Psychological review 110</source>
          ,
          <issue>2</issue>
          (
          <year>2003</year>
          ),
          <fpage>220</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>John</surname>
            <given-names>E</given-names>
          </string-name>
          <string-name>
            <surname>Hummel and David H Landy</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <article-title>From analogy to explanation: relaxing the 1: 1 mapping constraint very carefully</article-title>
          . In New Frontiers in Analogy Research: Proceedings of the Second International Conference on Analogy. New Bulgarian University,
          <fpage>211</fpage>
          -
          <lpage>221</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>John</surname>
            <given-names>E Hummel</given-names>
          </string-name>
          , John Licato, and
          <string-name>
            <given-names>Selmer</given-names>
            <surname>Bringsjord</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Analogy, explanation, and proof</article-title>
          .
          <source>Frontiers in human neuroscience 8</source>
          (
          <year>2014</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Paul</given-names>
            <surname>Kingsbury</surname>
          </string-name>
          and
          <string-name>
            <given-names>Martha</given-names>
            <surname>Palmer</surname>
          </string-name>
          .
          <year>2002</year>
          .
          <article-title>From TreeBank to PropBank</article-title>
          ..
          <source>In LREC. Citeseer.</source>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Hector J Levesque</surname>
            , Ernest Davis, and
            <given-names>Leora</given-names>
          </string-name>
          <string-name>
            <surname>Morgenstern</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>e Winograd schema challenge</article-title>
          ..
          <source>In AAAI Spring Symposium: Logical Formalizations of Commonsense Reasoning</source>
          , Vol.
          <volume>46</volume>
          . 47.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Hongjing</surname>
            <given-names>Lu</given-names>
          </string-name>
          , Alan L Yuille, Mimi Liljeholm, Patricia W Cheng, and
          <string-name>
            <surname>Keith</surname>
          </string-name>
          J Holyoak.
          <year>2008</year>
          .
          <article-title>Bayesian generic priors for causal learning</article-title>
          .
          <source>Psychological review 115</source>
          ,
          <issue>4</issue>
          (
          <year>2008</year>
          ),
          <fpage>955</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Christopher</surname>
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Manning</surname>
            , Mihai Surdeanu, John Bauer, Jenny Finkel,
            <given-names>Steven J.</given-names>
          </string-name>
          <string-name>
            <surname>Bethard</surname>
          </string-name>
          , and
          <string-name>
            <surname>David McClosky</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>e Stanford CoreNLP Natural Language Processing Toolkit</article-title>
          .
          <source>In Proceedings of 52nd Annual Meeting of the Association for Computational Linguistics: System Demonstrations</source>
          .
          <fpage>55</fpage>
          -
          <lpage>60</lpage>
          . h p://www.aclweb.org/anthology/P/P14/P14-5010
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Tomas</surname>
            <given-names>Mikolov</given-names>
          </string-name>
          , Ilya Sutskever, Kai Chen, Greg S Corrado, and
          <string-name>
            <given-names>Je</given-names>
            <surname>Dean</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Distributed representations of words and phrases and their compositionality</article-title>
          .
          <source>In Advances in neural information processing systems</source>
          .
          <volume>3111</volume>
          -
          <fpage>3119</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Nasrin</surname>
            <given-names>Mostafazadeh</given-names>
          </string-name>
          , Nathanael Chambers, Xiaodong He, Devi Parikh, Dhruv Batra, Lucy Vanderwende, Pushmeet Kohli, and
          <string-name>
            <given-names>James</given-names>
            <surname>Allen</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>A corpus and cloze evaluation for deeper understanding of commonsense stories</article-title>
          .
          <source>Proceedings of NAACL HLT</source>
          , San Diego, California, June. Association for Computational Linguistics (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>Nasrin</surname>
            <given-names>Mostafazadeh</given-names>
          </string-name>
          , Lucy Vanderwende, Wen-tau
          <string-name>
            <surname>Yih</surname>
            , Pushmeet Kohli, and
            <given-names>James</given-names>
          </string-name>
          <string-name>
            <surname>Allen</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Story Cloze Evaluator: Vector Space Representation Evaluation by Predicting What Happens Next</article-title>
          .
          <source>ACL</source>
          <year>2016</year>
          (
          <year>2016</year>
          ),
          <fpage>24</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <surname>Derek</surname>
            <given-names>C Penn</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Keith J Holyoak</surname>
          </string-name>
          , and
          <string-name>
            <surname>Daniel</surname>
          </string-name>
          J Povinelli.
          <year>2008</year>
          .
          <article-title>Darwin's mistake: Explaining the discontinuity between human and nonhuman minds</article-title>
          .
          <source>Behavioral and Brain Sciences</source>
          <volume>31</volume>
          ,
          <issue>02</issue>
          (
          <year>2008</year>
          ),
          <fpage>109</fpage>
          -
          <lpage>130</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>Karl</given-names>
            <surname>Picho</surname>
          </string-name>
          <article-title>a</article-title>
          and
          <string-name>
            <surname>Raymond J Mooney</surname>
          </string-name>
          .
          <year>2016</year>
          .
          <article-title>Statistical Script Learning with Recurrent Neural Networks</article-title>
          .
          <source>EMNLP</source>
          <year>2016</year>
          (
          <year>2016</year>
          ),
          <fpage>11</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>V.</given-names>
            <surname>Punyakanok</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Roth</surname>
          </string-name>
          , and
          <string-name>
            <given-names>W.</given-names>
            <surname>Yih</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>e Importance of Syntactic Parsing and Inference in Semantic Role Labeling</article-title>
          .
          <source>Computational Linguistics</source>
          <volume>34</volume>
          ,
          <issue>2</issue>
          (
          <year>2008</year>
          ). h p://cogcomp.cs.illinois.edu/papers/PunyakanokRoYi07.pdf
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>