<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>The Role of Pragmatics in Solving the Winograd Schema Challenge</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Adam Richard-Bollans</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Luc´ıa Go´ mez A´ lvarez</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Anthony G. Cohn</string-name>
          <email>a.g.cohng@leeds.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>School of Computing University of Leeds</institution>
          ,
          <addr-line>Leeds</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Different aspects and approaches to commonsense reasoning have been investigated in order to provide solutions for the Winograd Schema Challenge (WSC). The vast complexities of natural language processing (parsing, assigning word sense, integrating context, pragmatics and world-knowledge, ...) give broad appeal to systems based on statistical analysis of corpora. However, solutions based purely on learning from corpora are not currently able to capture the semantics underlying the WSC - which was intended to provide problems whose solution requires knowledge and reasoning, rather than statistical analysis of superficial lexical features. In this paper we consider the WSC as a means for highlighting challenges in the field of commonsense reasoning more generally. We begin by discussing issues with current approaches to the WSC. Following this we outline some key challenges faced, in particular highlighting the importance of dealing with pragmatics. We then argue for an alternative approach which favours the use of knowledge bases where the deep semantics of the different interpretations of commonsense terms are formalised. Furthermore, we suggest using heuristic approaches based on pragmatics to determine appropriate configurations of both reasonable interpretations of terms and necessary assumptions about the world.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        The Winograd Schema Challenge
        <xref ref-type="bibr" rid="ref19 ref33 ref41">(Levesque, Davis, and
Morgenstern 2012)</xref>
        was conceived as a new benchmark in
artificial intelligence, which would improve on the Turing Test
        <xref ref-type="bibr" rid="ref43">(Turing 1950)</xref>
        by removing the need for deception and
focusing more on understanding. The task is a particular type of
pronoun disambiguation problem. Sentences with a pronoun
and two candidate referents are given, and the task is to find
the correct referent of the pronoun. As the challenge is
intended to require genuine intelligence and understanding, the
sentences are supposed to be constructed in such a way that
syntactic constraints and semantic preference do not alone
enable the disambiguation. This construction is achieved in
part by finding pairs of sentences, differing only by one word
but where the pronoun reference is different. For example:
      </p>
      <sec id="sec-1-1">
        <title>The large ball crashed right through the table because</title>
        <p>it was made of [steel/styrofoam]. What was made of
[steel/styrofoam]? Answers: The ball/the table.1 (1)
The pronoun ‘it’ refers to either the ball or the table
depending on whether ‘steel’ or ‘styrofoam’ is used. In both
cases the syntactic structure remains the same and, supposing
that clear semantic preferences relating ‘steel’ and ‘crashing
through things’ or ‘styrofoam’ and ‘being crashed through’
cannot be easily learned from mining a large corpus, it is
hoped that any system which resolves the pronoun must use
some sort of genuine understanding.</p>
        <p>
          In the literature discussing the WSC and its motivation as
a benchmark we see example reasoning processes
incorporating detailed semantics of the language involved
          <xref ref-type="bibr" rid="ref19 ref20 ref29 ref33 ref37 ref41 ref8 ref9">(Davis 2013;
Levesque 2014; Levesque, Davis, and Morgenstern 2012;
Morgenstern and Ortiz Jr 2015)</xref>
          . This kind of approach
however has not been at the forefront of proposals to the
challenge. This is in large part due to the enormous complexity of
dealing with natural language and constructing large enough
knowledge bases to handle such varied contexts.
        </p>
        <p>
          In order to further the symbolic approach we investigate
some problems faced, mainly pragmatics. It is hoped that this
sort of analysis helps to shed light on what kind of reasoning
is needed where; and that heuristic methods will remove a
large portion of the burden of reasoning about natural
language. Along similar lines, a partial solution is provided in
          <xref ref-type="bibr" rid="ref38">(Schu¨ller 2014)</xref>
          , using relevance theory
          <xref ref-type="bibr" rid="ref34 ref42">(Sperber and Wilson
2004)</xref>
          to motivate selection of the best knowledge graph to
describe a sentence.
        </p>
        <p>In this paper we first explore what kind of reasoning
capabilities we expect a system to display when solving the
WSC and we analyse how some of the proposed approaches
compare to this. We then consider some key challenges for
solving the WSC using reasoning we consider appropriate;
in particular, that pragmatics and context are very difficult to
capture and semantics are hard to formalize due to vagueness.
Finally, we show how pragmatic considerations can help in
solving the WSC, specifically we consider how prototype
theory and heuristic methods can be used to support symbolic
approaches.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>What kind of reasoning are we looking for?</title>
      <p>
        We first consider the example above (1) given in
        <xref ref-type="bibr" rid="ref20">(Levesque
2014)</xref>
        , using the word ‘styrofoam’. Humans would
successfully resolve this by knowing particular properties of
sty1Taken from
www.cs.nyu.edu/faculty/davise/
papers/WinogradSchemas/WSCollection
rofoam, maybe some naive physics and even some general
properties of balls and tables.
      </p>
      <p>
        Levesque then considers what should be the outcome if
we change styrofoam to XYZZY, where XYZZY is some
material that we are given some facts about, one of the facts
being ‘It is ninety-eight percent air, making it lightweight and
buoyant’. Given this fact, humans would be able to reason
that the table is made of XYZZY. This is a part of intelligent
behaviour that we would like to replicate, and is clearly
dependent on having and being able to reason about detailed
knowledge. Further, it has been suggested as a possible
extension to the test to add a requirement for the solution to provide
a simple explanation of its choice
        <xref ref-type="bibr" rid="ref29 ref37 ref8">(Morgenstern and Ortiz Jr
2015)</xref>
        . This need for explanation would also seem to depend
on reasoning with detailed knowledge; in order to explain
why the table is made of styrofoam, it seems necessary to
have an understanding of the mechanics of the situation. The
ability to provide an explanation is also important more
generally for the field of commonsense reasoning, for example for
decision support systems that need to provide justifications
for decisions
        <xref ref-type="bibr" rid="ref14 ref16">(Hayes-Roth, Waterman, and Lenat 1984)</xref>
        .
      </p>
      <sec id="sec-2-1">
        <title>Versatile solutions</title>
        <p>
          The WSC was conceived as a new benchmark for artificial
intelligence; as such, we hope that solutions to the WSC will
provide tools for tackling a broader range of question
answering tasks and commonsense challenges. In this way, solutions
to the challenge should display versatility as well as making
advances in the WSC specifically, thus representing genuine
progress towards truly intelligent machines. Solutions which
are over-specific to the WSC and only provide insight into
this narrow set of coreference resolution problems are not
likely to be ‘engaging in behaviour that we would say shows
thinking in people’
          <xref ref-type="bibr" rid="ref19 ref33 ref41">(Levesque, Davis, and Morgenstern 2012)</xref>
          .
This is a similar but more general requirement than
elaboration tolerance
          <xref ref-type="bibr" rid="ref25">(McCarthy 1998)</xref>
          .
        </p>
        <p>
          The situations described in Winograd sentences (WS) are
generally common/normal occurrences; however, it is
desirable for AI systems to be able to reason about out-of-place
objects and strange scenarios. The ability to do this displays
a genuine understanding of what is going on. Levesque gives
the example ‘Can a crocodile run a steeplechase?’
          <xref ref-type="bibr" rid="ref20">(Levesque
2014)</xref>
          . Most humans would answer this easily using
basic knowledge about crocodiles (in particular that they
cannot jump) and what is necessary to be able to complete a
steeplechase. Of course, as noted by Levesque, a statistical
approach using the closed world assumption would be likely
to get the right answer to this question too as there is little
evidence of crocodiles running steeplechases. It would be less
likely however to answer the question correctly if the animal
was a gazelle (which presumably could run a steeplechase).
        </p>
        <p>Having briefly considered the kind of solutions we are
aiming for, we now look at how some existing approaches
compare to this.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Existing approaches to the challenge</title>
      <p>
        Since the inception of the WSC there has been some
theoretical discussion on the purpose of the challenge
        <xref ref-type="bibr" rid="ref20 ref29 ref37 ref8">(Davis
and Marcus 2015; Levesque 2014; 2017)</xref>
        , various methods
suggested for tackling the problem
        <xref ref-type="bibr" rid="ref1 ref29 ref31 ref33 ref37 ref38 ref39 ref41 ref8">(Sharma et al. 2015;
Schu¨ller 2014; Bailey et al. 2015; Rahman and Ng 2012;
Peng, Khashabi, and Roth 2015)</xref>
        , and four
implementations entered into the 2016 challenge2
        <xref ref-type="bibr" rid="ref15 ref23">(Liu et al. 2016;
Isaak and Michael 2016)</xref>
        (two of the competitors did not
release papers). The WSC is a particular type of anaphora
resolution task, on which there has been much work done in the
natural language processing community already
        <xref ref-type="bibr" rid="ref28 ref30 ref6">(Ng 2017;
Mitkov 2014; Carbonell and Brown 1988)</xref>
        ; however due to
the nature of the task, necessitating the use of world
knowledge, the methods employed are not wholly suitable for the
challenge.
      </p>
      <p>Formalizing the necessary aspects of reasoning to tackle
the WSC (spatial, temporal, causal, epistemic, ...) and
integrating them in one system is notoriously hard. Therefore, it
is not surprising that the space of genuine proposed solutions
is sparse, and that existing approaches are mostly based on
statistical methods, that circumvent the need for a precise
understanding of the semantics of the schemas by learning
likely answers from text corpora.</p>
      <p>In this section we analyse some of the solutions proposed
along these lines. We consider both their performance and
success on the challenge and also their achievements and
relevance for broader commonsense reasoning, which is the
ultimate aim of the WSC as a benchmark.</p>
      <sec id="sec-3-1">
        <title>Machine learning approaches</title>
        <p>
          Machine learning methods for anaphora resolution have been
used extensively over the past two decades
          <xref ref-type="bibr" rid="ref30">(Ng 2017)</xref>
          . In this
section we consider some of the best known such approaches
for tackling the WSC.
        </p>
        <p>
          The team that came first in the 2016 WSC challenge2 used
‘Commonsense Knowledge Enhanced Embeddings’
          <xref ref-type="bibr" rid="ref23">(Liu et al.
2016)</xref>
          which works by learning word representation vectors
from large text corpora while incorporating commonsense
knowledge as constraints in the training process. For the
competition the commonsense knowledge was obtained from
CauseCom — a set of cause and effect pairs such as ‘winning
causes happiness’
          <xref ref-type="bibr" rid="ref23">(Liu et al. 2016)</xref>
          — though the team has
also incorporated WordNet
          <xref ref-type="bibr" rid="ref27">(Miller 1995)</xref>
          and ConceptNet
          <xref ref-type="bibr" rid="ref33 ref41">(Speer and Havasi 2012)</xref>
          . A neural network is then trained
to answer yes or no when given candidate/pronouns pairs
(as vectors), and this network is then used to answer new
disambiguation problems.
        </p>
        <p>Though achieving a good performance on the challenge,
it would be down to chance whether it correctly answers the
XYZZY problem given by Levesque, whether it could be
used to solve the crocodile-steeplechase problem, or in future
how it could be developed to explain how it comes to the
conclusion.</p>
        <p>
          <xref ref-type="bibr" rid="ref33">Rahman and Ng (2012)</xref>
          have worked combining
multiple methods to resolve the pronoun for a large corpus of
WSs. This work achieved high results on their corpus, 73.1%.
However, the corpus selection has been criticized for
containing redundancy
          <xref ref-type="bibr" rid="ref40">(Sharma 2014)</xref>
          . Further, the approach relies
2www.cs.nyu.edu/faculty/davise/papers/
WinogradSchemas/WS.html
heavily on statistical methods for assessing the semantic
preferences of types and events e.g. a lion is a type of predator
and being the subject of a kill event makes one more likely
to be the object of an arrest event. It is clear that ‘lions eat
zebras because they are predators’ is not a ‘Google-proof’
WS and should be discarded. When such type distinctions
are not useful, the system may rely on FrameNet
          <xref ref-type="bibr" rid="ref2">(Baker,
Fillmore, and Lowe 1998)</xref>
          ; in the case of ‘John killed Jim
so he was arrested’, FrameNet gives John the role of ‘killer’
and Jim the role of ‘victim’ and the system, using statistical
methods, concludes that it is more likely for a ‘killer’ (John)
to be arrested. In this case the system resolves the pronoun
successfully. However, this takes no account of the
importance of the connective: changing the sentence to ‘John killed
Jim after he was arrested’ should force one to re-evaluate the
disambiguation.
        </p>
        <p>
          Work by
          <xref ref-type="bibr" rid="ref31">Peng et al. (2015)</xref>
          has been successful, achieving
higher results (76.4%) than Rahman and Ng on the same
corpus. The technique is similar to the FrameNet approach of
          <xref ref-type="bibr" rid="ref33 ref41">(Rahman and Ng 2012)</xref>
          but they also take connectives into
account. This approach can give crude, and clearly problematic,
forms of knowledge such as ‘fflower has polleng is more
likely than fbee has polleng’; to more reasonable knowledge
such as ‘the subject of “be afraid of ” is more likely than the
object of “be afraid of ” to be the subject of “get scared of ”’.
Though these sorts of techniques will likely prove very useful
for natural language processing, and may even manage to
pass the WSC, there is a fundamental issue that these
techniques are learning about the likelihood of combinations of
words in corpora and there appears to be little in the way of
transferable knowledge or understanding. For example, it is
clear that the kind of background knowledge necessary to
solve the crocodile-steeplechase problem is not present.
        </p>
        <p>Rather than applying reasoning to knowledge, these
techniques are geared towards mining what we may call
commonsense rules. We discuss the nature of such rules in the
following section.</p>
      </sec>
      <sec id="sec-3-2">
        <title>Commonsense rules</title>
        <p>It is clear that, in the WSC, it appears possible to resolve
pronoun ambiguity through an appeal to normality — heavy
things cannot be lifted, younger people are fitter, useless
objects go in the bin while useful tools are kept in storage
etc... Hence, a large part of the suggested approaches to the
WSC have been about ways of finding and/or incorporating
such ‘commonsense rules’. We believe, however, that this is
a rather crude view of commonsense reasoning and outline
some problems of these approaches below.</p>
        <p>
          One proposed approach is that we reduce some of the
implied causation in WSs to correlation
          <xref ref-type="bibr" rid="ref1">(Bailey et al. 2015)</xref>
          .
This uses ‘correlation formulas’ of the form F G, such
as ‘fit into(x; y) large(y)’ to say that ‘stuff fitting into y’
is correlated with ‘y being large’. Some inference rules are
given governing such correlation formulas and it is shown
how these could be used to justify a solution to a WS. This
approach is however problematic. It is analogous to a discussion
in
          <xref ref-type="bibr" rid="ref5">(Bunt and Black 2000)</xref>
          — by reducing to mere convention
the reason why ‘There is a howling gale in here!’ is
understood as a command to close the window, we are
oversimplifying and missing out more important reasoning processes,
including context. Similarly, if we were to find a list of
commonsense correlations like ‘fit into(x; y) large(y)’ through
corpus mining, we are ripping the words out of context and
may be missing out important reasoning processes.
        </p>
        <p>
          This is not to say that conventions do not exist or form an
important part of commonsense reasoning. Natural language
is full of conventions that we may rely upon to communicate.
For example, considering the sentence ‘Sam chopped down
the tree’ there is a default assumption that the chopping is
done with an axe. This kind of convention can be considered
as part of linguistic knowledge
          <xref ref-type="bibr" rid="ref32">(Pustejovsky 1991)</xref>
          . However,
reasoning based solely on conventions may be too crude, as it
does not take contextual factors into consideration. Say that
we know that Sam is holding a sword, then we may reject
the default assumption that Sam chops down the tree with
an axe. One way of dealing with the context dependency of
such conventions may be to apply context frames, as in
          <xref ref-type="bibr" rid="ref24">(McCarthy 1993)</xref>
          , i.e. in the context of Sam holding a sword, the
statement ‘Sam chopped down the tree’ suggests that Sam
did the chopping with a sword rather than an axe. However,
even if we can create appropriate context frames using salient
aspects of context, it seems that the process of creating
convention/context pairs would continue ad infinitum. We would
hope that reasoning removes the necessity for a lot of these
rules e.g. when someone is holding an appropriate tool, T, for
performing action, A, and we are told that they performed
action A, then we can assume that they have used T to do A.
        </p>
        <p>
          The tactic for many approaches is to begin by learning
commonsense knowledge from large text corpora or by
integrating natural language knowledge bases. Part of the appeal
of this is that knowledge can be exploited without having to
translate between formal and natural language. However, the
methods for extracting commonsense knowledge from the
Web can be problematic. Language is used in an efficient way
and commonsense knowledge is often left implicit
          <xref ref-type="bibr" rid="ref29 ref37 ref8">(Schu¨ller
and Kazmi 2015)</xref>
          .
        </p>
        <p>Even if we were able to overcome some of the problems of
mining commonsense, do we want to use reasoning that relies
solely on these correlations and rules? Though they may be
helpful for certain applications, the reasoning mechanisms
need to incorporate less crude knowledge. Regarding the
desire for versatility and considering some of the problems
listed on the Common Sense Problem Page3, it is clear that
this approach is over-specific to the WSC. It would also
clearly be hard to mine relations between crocodiles and
steeplechases in this way! Moreover, any explanation of the
disambiguation given by such a system would not be very
enlightening. Considering schema (1) with ‘steel’; explaining
why ‘it’ refers to the ball by saying that ‘steel things are more
likely to crash through things than to be crashed through’ is
not a reasonable explanation. Even the ability to cite a salient
property of steel like ‘steel is hard’ would be an important
improvement.</p>
        <p>
          The approaches outlined above at best only incorporate
shallow semantic features and do not appear to exhibit the
3www-formal.stanford.edu/leora/
commonsense/
kind of intelligent behaviour the challenge was designed to
test. We believe that, in order to carry out complex
inferences and really understand the world, some definitions of
the natural language in terms of more refined primitives is
often necessary. It is necessary to have genuine world
knowledge of entities, as well as their physical, social/historical
and functional attributes, as in
          <xref ref-type="bibr" rid="ref4">(Bennett 2005)</xref>
          , and be able
to reason about that knowledge, e.g. crocodiles have short
legs and long bodies, making them unsuitable candidates
for a steeplechase, rather than superficial knowledge about
relations between entities which are mined from corpora, e.g.
crocodiles do not run steeplechases. A line may be drawn
by the distinction between reasoning from first principles
and reasoning by analogy. They can both be valid forms of
reasoning, but reasoning by analogy alone is not enough to
be considered intelligent.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Key challenges</title>
      <p>This section outlines some particular problems that need
resolving in order to tackle the WSC and for commonsense
reasoning systems more generally.</p>
      <sec id="sec-4-1">
        <title>Pragmatics</title>
        <p>
          A large part of the complexity of the WSC comes from
pragmatic considerations. There are varying positions on the
definition of pragmatics
          <xref ref-type="bibr" rid="ref7">(Carston 1999)</xref>
          , however it is generally
understood as the field concerned with extra-linguistic
factors, such as context, and how they allow the understanding
of a speaker’s intended meaning.
        </p>
        <p>
          Semantic considerations are clearly essential but they are
generally not enough in order to reach a conclusion about
the disambiguation for a WS. This is an example of semantic
underdeterminacy — that from only considering the literal
meanings of terms in a sentence and not accounting for the
intended meaning, we do not obtain a truth-evaluable
proposition. For example, the sentence ‘Tom threw his school bag
down to Ray after he reached the top of the stairs’ does not
contain much information if we only consider the semantics.
We also need to consider the intention of the speaker and we
may infer this from the decisions the speaker takes regarding
the specific choice of language, what information is omitted,
what is left ambiguous, the phrasing of the sentence etc...
Indeed, Kempson argues that ‘the articulation of semantics
[does not alone] provide the full propositional content/logical
form/truth conditions expressed by a sentence’
          <xref ref-type="bibr" rid="ref16">(Kempson
1984)</xref>
          .
        </p>
        <p>To evidence this view, we can see that for many WSs
wrongly disambiguating the pronoun does not necessarily
violate world knowledge. For example, when dealing with
the sentence:</p>
        <p>The trophy does not fit into the suitcase because itx is
too large1 (2)
there are various interpretations of ‘large’ which give no
definite disambiguation. If we imagine a trophy and suitcase
to be vase-shaped, with a wide base, narrow stem and wide
top, and that the trophy fits into the suitcase, it is possible that
making the suitcase larger via a scale projection would make
the trophy no longer fit. It is in part by making pragmatic
considerations that we can assign appropriate interpretations
to these terms and thus disambiguate the pronoun.</p>
        <p>
          Moreover, even in the sentences where each term can be
precisely and appropriately defined we can still have
semantic underdeterminacy. Is it often the case that an utterance is
not totally explicit and leaves the reader to fill in the gaps
with available assumptions and inferences
          <xref ref-type="bibr" rid="ref7">(Carston 1999)</xref>
          .
One of the ways that a hearer may fill in these gaps and
infer a speaker’s intention is by assuming Grice’s Maxims for
co-operative communication
          <xref ref-type="bibr" rid="ref12">(Grice 1975)</xref>
          ; e.g. the ‘Quantity
Maxim’, stating: ‘Make your contribution as informative as is
required’ and ‘Do not make your contribution more
informative than is required’. So for instance, if a speaker goes into a
lot of detail when making an utterance, we may assume that
there is particular reason for this and can infer things based
on this knowledge. This kind of pragmatic inference is also
important for written text, and hence the WSC. Therefore, as
it stands, any solution to the WSC needs some mechanisms
for coping with this implicit knowledge.
        </p>
        <p>In the next section we consider some particular examples
of this sort of inference when addressing a WS.</p>
      </sec>
      <sec id="sec-4-2">
        <title>Assumptions about the world</title>
        <p>When facing any WS there are multiple commonsense
principles that apply which allow us to create an accurate model of
the situation. What we aim to achieve is some guidance on
how to choose these principles and when they apply. To this
end we examine the following WS:</p>
        <sec id="sec-4-2-1">
          <title>Tom threw his school bag down to Ray after hex</title>
          <p>reached the [top/bottom] of the stairs. Who reached the
[top/bottom] of the stairs? Answer: top: Tom. bottom:
Ray.1 (3)
We will use this example to help elucidate some of the
complexities faced, including the initial position of objects
and relevant objects.</p>
          <p>The main idea of this sentence is that to throw something
down to someone, that person must be below you. We then
use the idea of what it means to be at the top of something,
i.e. that if Ray is at the top of the stairs then he cannot be
below Tom. This is however not as clear as it seems.
Initial position It is possible that Tom is on some balcony
above the stairs and waits for Ray to reach the top of the stairs
before throwing the bag down to Ray. So why do we like
the answer ‘Tom’? It appears we assume that Tom and Ray
are initially in a similar location, or to be more precise, that
they both have the same relation to any given landmark — in
this case the stairs. Character x reaching the top of the stairs
implies that x has moved upwards. Not given any information
on the other character, y, we assume they have not moved
and so x is likely to be above y.</p>
          <p>Alternatively, x may have been walking along a corridor
to reach the top of the stairs. In this scenario we have two
locations to consider, the corridor and the stairs. We suppose
that Tom and Ray are on the stairs or in the corridor. In this
case it would make no sense for Ray to be at the top of the
stairs, as then Tom would not be able to throw anything down
to him (from the corridor or the stairs); so we suppose that it
must be Tom who walks along the corridor to reach the top
of the stairs and throw the bag down to Tom.</p>
          <p>We appeal to a rule that in some narrative, unless we have
reason to infer otherwise, characters are nearby/in the same
place. This idea can be explained by Grice’s quantity maxim
i.e. there is no pertinent difference in the positions of either
Tom or Ray; if there were then the quantity maxim says it
should be made known.</p>
          <p>This rule however does not always hold. Imagine we
replace ‘stairs’ with ‘swimming pool’:</p>
        </sec>
        <sec id="sec-4-2-2">
          <title>Tom threw his school bag down to Ray after hex reached</title>
          <p>the top of the swimming pool. Who reached the top of
the swimming pool? Answer: Ray.</p>
          <p>In this scenario x reaches the top of the swimming pool,
breaking the surface of the water. x is then not in a position
to throw something like a school bag downwards, as it is
pretty hard to throw textile objects through water. Hence, we
imagine that x is not Tom, but Ray, and that Tom must be
stood somewhere above the swimming pool.</p>
          <p>Relevant objects In general in the WSC to come to a
conclusion we only need to reason about entities that are
explicitly mentioned. In the school bag example we reason about
the two characters in the narrative, Tom and Ray, the staircase
and the school bag itself. Combining knowledge of actions
like ‘throwing’ ‘reaching the top of’ etc.. with knowledge of
these objects. In general then, we do not need to appeal to
the existence of extra entities in order to come to a
conclusion. This can also be explained by the quantity maxim, the
sentence should provide the necessary objects for the reader
to make sense of the sentence.</p>
          <p>
            However, as previously discussed, certain words or
phrasings indirectly suggest the existence of certain entities, as
in the ‘Sam chopped down the tree’ example. We can in
part account for these entities by encoding into a lexicon
            <xref ref-type="bibr" rid="ref32">(Pustejovsky 1991)</xref>
            , though these are conventions that will
not always hold. Therefore a defeasible reasoning process is
necessary to select the most appropriate interpretation.
          </p>
          <p>To conclude our discussion about assumptions about the
world, we see that appropriate assumptions need to be made
in order to reach the right conclusion. Further, we believe that,
to varying extents, these kinds of considerations arise when
analysing most WSs appearing in the collection maintained
by Davis1. However, the assumptions are dependent on the
specific situation and we need to discern somehow when the
assumptions are appropriate. Deciding when to accept these
assumptions should include pragmatic considerations. For
example, it is lexical and semantic knowledge that suggest
the existence of an axe in the sentence ‘Sam chopped down
the tree’, however it is a pragmatic task to actually infer
this. This motivates a heuristic process which incorporates
pragmatics and gives preference to default assumptions, we
will discuss this idea later.</p>
        </sec>
      </sec>
      <sec id="sec-4-3">
        <title>Formalizing commonsense knowledge: level of detail and vagueness</title>
        <p>
          An important issue is to recognize the level of semantics that
one believes is appropriate for a solution to the WSC. Our
discussion so far has motivated a detailed level of knowledge.
Further, there is evidence that, even for coreference
problems that would be considered easy with respect to the WSC,
incorporating shallow semantic features is not enough
          <xref ref-type="bibr" rid="ref10">(Durrett and Klein 2013)</xref>
          . Yet, if we are to solve the WSC using
deeper semantics, it is clear that the necessary commonsense
knowledge would involve the formalization of a notoriously
extensive knowledge base. How to obtain and organize such
a large knowledge base is unclear.
        </p>
        <p>
          On the one hand, due to the variety and scope necessary,
mining commonsense knowledge is appealing; however, as
previously discussed, the available methods and nature of
text corpora pose limitations to obtaining deep knowledge,
which is complex and commonly not explicit. On the other
hand, hand crafted knowledge bases such as CYC
          <xref ref-type="bibr" rid="ref17">(Lenat
1995)</xref>
          , which incorporate a deeper level of knowledge, have
had limited success and it is not clear how they should be
exploited.
        </p>
        <p>
          Beyond the problem of its acquisition, it is well known that
commonsense knowledge is hard to formalize, particularly if
the required level of detail involves the semantics of natural
terms to be preserved. Vagueness and ambiguity are inherent
to natural language and, for that reason, it is problematic
to prescribe single strict interpretations to natural terms. To
illustrate this, consider the WS (3) and imagine the case of
a naive definition of a relation at the top of (x; y) x is on
y and for any z which is part of y, x is not below z. We see
that this fails for multiple reasons.
1. If Tom were one step below the very last one, it could still
be considered that he is at the top of the stairs, particularly
if Ray were well below him. We call it sorites vagueness
when there is a the lack of a clear threshold of application
of a term.
2. If we change ‘stairs’ to ‘building’ we might say that
Tom is at the top of a building because he is on the top
floor, rather than on the roof. In that case we are
shifting the interpretation of the predicate to something like
at the top of (x; y) z is the top part of y and x is on
z. There may also be many admissible interpretations of
what it means for z to be the top part of y. We call the
multiplicity of conceptually distinct interpretations of
natural terms conceptual vagueness. Further discussion on
the multiple interpretations of natural language terms and
their role in knowledge bases and ontologies can be found
in
          <xref ref-type="bibr" rid="ref4">(Bennett 2005)</xref>
          .
        </p>
        <p>
          Much of the work done in acquiring commonsense
knowledge circumvents vagueness in different ways, such as
using shallow semantics or microtheories that do not need to
be consistent with one another. Various theories, however,
have been proposed for dealing with vagueness. Fuzzy logic
          <xref ref-type="bibr" rid="ref45">(Zadeh 1965)</xref>
          stands as an intuitive solution for modelling
sorites vagueness by assigning degrees of truth. More
interesting for this research, supervaluation semantics
          <xref ref-type="bibr" rid="ref11">(Fine
1975)</xref>
          is based on the idea that vague language can be
interpreted in many different precise ways, each of which can be
logically conceptualised in a precisification
          <xref ref-type="bibr" rid="ref13 ref3">(Bennett 2001;
Go´mez A´ lvarez and Bennett 2017)</xref>
          , thus also offering support
for modelling conceptual vagueness.
        </p>
        <p>So where do all these considerations lead us? In order to
reach the kind of solution we desire, we must be able to deal
with semantic underdeterminacy — part of which involves
deciding when to use appropriate commonsense assumptions
— and also make use of a vast amount of detailed knowledge
while dealing with the associated problems of vagueness.</p>
        <p>With these issues in mind, we now consider some avenues
for further work.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>The role of pragmatics in solving the WSC</title>
      <p>In the previous sections we have highlighted how current
approaches, regardless of their success in solving schemas,
have provided limited support for the kind of intelligent
behaviour that we would like to replicate. Here, in an attempt
to account for some of the key challenges, we propose an
alternative approach, favouring the use of knowledge bases
where the deep semantics of the different interpretations of
commonsense terms are formalised. Furthermore, we suggest
using heuristic approaches based on pragmatics to determine,
in the context of each particular schema, appropriate
configurations of both reasonable interpretations of the terms and
necessary assumptions about the world.</p>
      <p>For this purpose we first motivate the use of prototypes
for categories and relations and then develop how heuristic
methods can provide a manageable way of using pragmatic
knowledge for the disambiguation of WSs.</p>
      <sec id="sec-5-1">
        <title>Appealing to prototypicality</title>
        <p>
          There is various work in pragmatics and cognitive science
highlighting the importance of using prototypes: in utterance
interpretation defaults are assigned before contextual and
pragmatic considerations are taken into account
          <xref ref-type="bibr" rid="ref22 ref34">(Levinson
1995; Recanati 2004)</xref>
          and there is also evidence for the human
preference for good examples (prototypes) of some category
as opposed to boundary cases and, further, that prototypes are
associated with the least processing effort
          <xref ref-type="bibr" rid="ref36">(Rosch 1978)</xref>
          . In
the particular scenario of a WS, we argue that the way vague
terms are presented leads the reader to interpret them
considering prototypical instances fitting the described scenario.
For instance, when one reasons about the WS (2) involving
the trophy and the suitcase, it is not necessary to worry about
a precise semantic commitment for the notion of larger, but
instead to evaluate the sentence considering clear cases that
satisfy most of the possible interpretations.
        </p>
        <p>
          Some of the previously discussed approaches work along
similar lines, using general commonsense rules and a notion
of correlation which appeal to a sense of typicality. However,
we believe that this should be more nuanced and that the deep
semantics of different interpretations should be preserved.
Hence, we propose an approach using ideas from prototype
theory
          <xref ref-type="bibr" rid="ref11 ref12 ref35">(Rosch and Mervis 1975)</xref>
          to differentiate prototypical
instances of vague terms and relations from borderline cases
within a supervaluationist approach.
        </p>
        <p>
          Much work has been done on how to pinpoint prototypical
members of categories, mainly using vector analysis or
conceptual spaces to find the centroid of a concept
          <xref ref-type="bibr" rid="ref18 ref44">(Verheyen,
Ameel, and Storms 2007; Lenci 2011)</xref>
          . However, it is not
clear how one could reason with this to resolve a WS, and
further, we are not only interested in picking a prototypical
example from a category, say from the class ‘pet’ or ‘things
that we eat’. Instead, we would also like to find prototypical
instances of relations that can be used to compare an infinite
number of objects. Although there is some work done on
vector analysis for relationships between words
          <xref ref-type="bibr" rid="ref10 ref26">(Mikolov,
Yih, and Zweig 2013)</xref>
          , in particular for analogy problems,
it does not appear to be applicable to this sort of reasoning
problem.
        </p>
        <p>
          Suppose we have a vague term, like ‘smaller’. How can we
decide on prototypical instances of this relation? Adopting
the supervaluation approach we would have a collection of
precise interpretations of its meaning. Following motivation
from
          <xref ref-type="bibr" rid="ref11 ref12 ref35">(Rosch and Mervis 1975)</xref>
          — considering shared
properties of classes — in an ideal scenario prototypical instances
of ‘smaller’ share properties across all instances of ‘smaller’
i.e. a prototypical instance of smaller is considered smaller
in all plausible interpretations. Consider the definitions for
‘smaller’ given in
          <xref ref-type="bibr" rid="ref9">(Davis 2013)</xref>
          :
        </p>
        <sec id="sec-5-1-1">
          <title>1. Smaller(a; b)</title>
        </sec>
        <sec id="sec-5-1-2">
          <title>2. Smaller(a; b)</title>
        </sec>
        <sec id="sec-5-1-3">
          <title>3. Smaller(a; b)</title>
        </sec>
        <sec id="sec-5-1-4">
          <title>4. Smaller(a; b)</title>
        </sec>
        <sec id="sec-5-1-5">
          <title>VolumeOf (a) &lt; VolumeOf (b)</title>
        </sec>
        <sec id="sec-5-1-6">
          <title>DiameterOf (a) &lt; DiameterOf (b)</title>
          <p>a</p>
          <p>b
9s(s &gt; 1 ^ b = Scale(a; s))</p>
          <p>In this scenario, there are certainly pairs of objects that fall
into all four categories (e.g. a sphere of radius 1 is smaller
than a sphere of radius 2 in all the above senses). Hence,
it would be appropriate to take the conjunction of all four
definitions as a requirement for an instance to be considered a
prototypical case of ‘smaller’. However, in certain scenarios
it may be inappropriate to take the conjunction in this way,
as some definitions may be conflicting. In this case different
metrics can be proposed for selecting prototypes that satisfy
most of the interpretations.</p>
          <p>Finally, our main claim in this section is twofold. On the
one hand, we consider that an understanding of typicality
is necessary for commonsense reasoning — that by default
we should consider prototypes. On the other hand, a process
which can only reason over prototypical definitions is clearly
flawed in many respects as it creates over-simplification.
Humans often use context to help narrow definitions, for
example defining ‘smaller’ in a particular way makes sense
when talking about ‘fitting in’. Hence we believe that a good
approach should reflect the diversity of possible
interpretations of vague terms and that an engine based on pragmatics
should guide the selection of appropriate alternatives when
the prototype is not suitable.</p>
        </sec>
      </sec>
      <sec id="sec-5-2">
        <title>Heuristics standing in for pragmatics</title>
        <p>
          In this paper we have discussed some approaches proposed
for the WSC relying on heuristic methods in different ways
          <xref ref-type="bibr" rid="ref23 ref29 ref31 ref33 ref37 ref41 ref8">(Rahman and Ng 2012; Peng, Khashabi, and Roth 2015;
Liu et al. 2016)</xref>
          . Overall, we concluded that heuristics do
not provide satisfactory solutions when reduced to evaluating
shallow semantic notions such as correlation.
        </p>
        <p>Instead, as has been argued, we believe that a good solution
to the WSC should disambiguate the pronoun by considering
the most plausible configuration of the scenario described,
and the process of finding it should incorporate rich syntactic,
semantic and pragmatic considerations. However, although
advocating deeper semantics and symbolic based approaches
that allow for the kind of reasoning that we want (see section
above), we propose that heuristic methods have a key role
in the WS resolution: that of simplifying the space of
possibilities and estimating reasonably good configurations of
precisifications and necessary assumptions about the world.</p>
        <p>
          As we have highlighted above in order to carry out
satisfactory reasoning we believe a system should give preference
to both commonsense assumptions about the world as well as
prototypical interpretations of the terms involved. These
however should only be preferences rather than concrete rules.
When to accept or reject these default assumptions requires
knowledge and pragmatic understanding. The ability for this
complex mix of pragmatics and world knowledge to
contradict itself means that possible solutions or configurations
of a described scenario are not unique. For example, when
discussing the issue of throwing a school bag in a
swimming pool above, the implausibility of throwing a school
bag through water outweighed the assumption of Tom and
Ray being in the same place. However, we may also consider
that the assumption of characters being in the same place
outweighs the usual interpretation of ‘throw down’ and ‘top’:
supposing Tom and Ray are both stood in the swimming pool,
we may interpret ‘throw down’ as ‘throw horizontally away
from the end of the swimming pool’ and ‘top of the
swimming pool’ to denote the end of the swimming pool. The
result would then be to disambiguate the pronoun as ‘Tom’
rather than ‘Ray’. This second interpretation is not wrong,
however when ‘throw down’ and ‘top’ are interpreted in their
usual way there is a plausible inference that Tom and Ray
are not both located in the swimming pool. This would then
be an example of a ‘conversational implicature’
          <xref ref-type="bibr" rid="ref12">(Grice 1975)</xref>
          and explain why the writer of the sentence did not explicitly
give Tom and Ray’s initial locations. Hence in the first
interpretation we have a good explanation for violating the default
that Tom and Ray are located in the same place and we also
interpret all the terms in a usual fashion, therefore making
this interpretation appear to be the valid one.
        </p>
        <p>
          Being able to leverage these kinds of inferences is an
important and difficult task in commonsense reasoning. Along
these lines, one avenue
          <xref ref-type="bibr" rid="ref38">(Schu¨ller 2014)</xref>
          adopted in tackling
the WSC has been to explore relevance theory
          <xref ref-type="bibr" rid="ref34 ref42">(Sperber and
Wilson 2004)</xref>
          . This theory, inspired by Grice’s work, is based
on the idea that an utterance can have a variety of
interpretations, and that it is through parsing, disambiguating terms,
resolving pronouns and adding pragmatic inference as well
as appropriate assumptions based on context that one can
comprehend the meaning of an utterance. The principle
guiding these tasks is the idea of maximizing relevance4. Schu¨ller
uses these ideas to motivate a heuristic process for reasoning
over graphs, where a fitness function is employed to find
relevant combinations that provide a disambiguation. Moreover,
the resulting graph can be read off to get some idea of how
4An input is said to be relevant if a worthwhile conclusion is
drawn from it. An input is more relevant if it yields a greater positive
cognitive effect for less processing effort.
it came to that disambiguation, potentially satisfying
Morgenstern and Ortiz’s requirement of a simple explanation. In
spite of being preliminary research, in our view its reasonable
results suggest that fruitful work can be done in further
developing heuristic methods to assess the pragmatic and semantic
considerations that govern reasonable disambiguations of
natural language.
        </p>
        <p>To conclude this section, it is our claim that this use of
heuristics is much more in keeping with the nature of the
WSC. That what should be simplified in order to keep the
task manageable is not so much the deep semantics of natural
terms, but the process of selecting and integrating relevant
interpretations and background knowledge in the particular
context of the resolution of each sentence.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Conclusion</title>
      <p>In this paper we have discussed the nature of the WSC as
a benchmark, highlighting the shortcomings of several
current approaches and providing motivation for a more detailed
level of knowledge. We have also analysed some of what we
consider to be key challenges, in particular drawing attention
to the need to take account of pragmatic considerations. To
begin addressing these challenges, we have suggested using
frameworks able to support the detailed semantics of natural
terms while accounting for its vagueness. Moreover, that their
complexity can be manageable with the use of prototypes,
which should be identified and used by default, and, finally,
that heuristic methods can be used to incorporate varying
semantic interpretations as well as assumptions about the
world, which maintain the pragmatic principles of
cooperative communication.</p>
      <p>In conclusion, it is our view that, while heuristic
mechanisms are necessary to deal with natural language and to
reduce the complexity of commonsense reasoning, they should
not be used to over-simplify the semantics of natural terms.
Instead, we believe that applications along the lines of
theoretical studies in pragmatics can play a significant role in
the selection of good interpretations of natural terms and
to enrich the provided descriptions of the world with the
appropriate implicit knowledge.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgements</title>
      <p>Thanks to Brandon Bennett for helpful discussion and to the
anonymous reviewers for their useful feedback.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Bailey</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Harrison</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Lierler</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Lifschitz</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ; and Michael,
          <string-name>
            <surname>J.</surname>
          </string-name>
          <year>2015</year>
          .
          <article-title>The Winograd Schema Challenge and Reasoning about Correlation</article-title>
          .
          <source>In Working Notes of the Symposium on Logical Formalizations of Commonsense Reasoning.</source>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Baker</surname>
            ,
            <given-names>C. F.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Fillmore</surname>
            ,
            <given-names>C. J.;</given-names>
          </string-name>
          and
          <string-name>
            <surname>Lowe</surname>
            ,
            <given-names>J. B.</given-names>
          </string-name>
          <year>1998</year>
          .
          <article-title>The berkeley framenet project</article-title>
          .
          <source>In Proceedings of COLING/ACL</source>
          ,
          <fpage>86</fpage>
          -
          <lpage>90</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>Bennett</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <year>2001</year>
          .
          <article-title>What is a Forest? On the vagueness of certain geographic concepts</article-title>
          .
          <source>Topoi</source>
          <volume>20</volume>
          (
          <issue>2</issue>
          ):
          <fpage>189</fpage>
          -
          <lpage>201</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Bennett</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <year>2005</year>
          .
          <article-title>Modes of concept definition and varieties of vagueness</article-title>
          .
          <source>Applied Ontology</source>
          <volume>1</volume>
          (
          <issue>1</issue>
          ):
          <fpage>17</fpage>
          -
          <lpage>26</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Bunt</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Black</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          <year>2000</year>
          .
          <article-title>The ABC of Computational Pragmatics</article-title>
          . In Bunt, H., and
          <string-name>
            <surname>Black</surname>
          </string-name>
          , W., eds.,
          <source>Natural Language Processing</source>
          , volume
          <volume>1</volume>
          . Amsterdam: John Benjamins Publishing Company.
          <fpage>1</fpage>
          -
          <lpage>46</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Carbonell</surname>
          </string-name>
          , J. G.,
          <article-title>and</article-title>
          <string-name>
            <surname>Brown</surname>
          </string-name>
          , R. D.
          <year>1988</year>
          .
          <article-title>Anaphora resolution: a multi-strategy approach</article-title>
          .
          <source>In Proceedings of the 12th Conference on Computational linguistics</source>
          , volume
          <volume>1</volume>
          ,
          <fpage>96</fpage>
          -
          <lpage>101</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Carston</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <year>1999</year>
          .
          <article-title>The semantics/pragmatics distinction: A view from relevance theory</article-title>
          . In Turner, K., ed.,
          <article-title>The semantics/pragmatics interface from different points of view</article-title>
          . Oxford, UK: Elsevier.
          <fpage>85</fpage>
          -
          <lpage>125</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Davis</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Marcus</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <year>2015</year>
          .
          <article-title>Commonsense reasoning and commonsense knowledge in Artificial Intelligence</article-title>
          .
          <source>Communications of the ACM</source>
          <volume>58</volume>
          (
          <issue>9</issue>
          ):
          <fpage>92</fpage>
          -
          <lpage>103</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Davis</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <year>2013</year>
          .
          <article-title>Qualitative Spatial Reasoning in Interpreting Text and Narrative</article-title>
          .
          <source>Spatial Cognition &amp; Computation</source>
          <volume>13</volume>
          (
          <issue>4</issue>
          ):
          <fpage>264</fpage>
          -
          <lpage>294</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>Durrett</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Klein</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <year>2013</year>
          .
          <article-title>Easy Victories and Uphill Battles in Coreference Resolution</article-title>
          .
          <source>In EMNLP</source>
          ,
          <year>1971</year>
          -
          <fpage>1982</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>Fine</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <year>1975</year>
          .
          <article-title>Vagueness, truth and logic</article-title>
          .
          <source>Synthese</source>
          <volume>30</volume>
          (
          <issue>3</issue>
          ):
          <fpage>265</fpage>
          -
          <lpage>300</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <surname>Grice</surname>
            ,
            <given-names>H. P.</given-names>
          </string-name>
          <year>1975</year>
          .
          <article-title>Logic and conversation</article-title>
          .
          <source>In Syntax and Semantics</source>
          , Vol.
          <volume>3</volume>
          ,
          <string-name>
            <given-names>Speech</given-names>
            <surname>Acts</surname>
          </string-name>
          . New York: Academic Press.
          <fpage>41</fpage>
          -
          <lpage>58</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <article-title>Go´mez A´lvarez</article-title>
          , L., and
          <string-name>
            <surname>Bennett</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>Classification, Individuation and Demarcation of Forests: formalising the multi-faceted semantics of geographic terms</article-title>
          .
          <source>In 13th International Conference on Spatial Information Theory. Leibniz International Proceedings in Informatics.</source>
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <surname>Hayes-Roth</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Waterman</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Lenat</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <year>1984</year>
          .
          <article-title>Building expert systems</article-title>
          . Reading, MA: Addison-Wesley.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <surname>Isaak</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Michael</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <year>2016</year>
          .
          <article-title>Tackling the Winograd Schema Challenge Through Machine Logical Inferences</article-title>
          . In Pearce, D., and Sofia Pinto, H., eds.,
          <source>STAIRS</source>
          , volume
          <volume>284</volume>
          <source>of Frontiers in Artificial Intelligence and Applications</source>
          . IOS Press.
          <volume>75</volume>
          -
          <fpage>86</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <surname>Kempson</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <year>1984</year>
          .
          <article-title>Pragmatics, anaphora and logical form</article-title>
          . In Schriffin, D., ed., Meaning, form
          <article-title>and use in context: linguistic applications</article-title>
          . Washington, DC: Georgetown University Press. 1-
          <fpage>10</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <surname>Lenat</surname>
            ,
            <given-names>D. B.</given-names>
          </string-name>
          <year>1995</year>
          .
          <article-title>CYC: A large-scale investment in knowledge infrastructure</article-title>
          .
          <source>Communications of the ACM</source>
          <volume>38</volume>
          (
          <issue>11</issue>
          ):
          <fpage>33</fpage>
          -
          <lpage>38</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <surname>Lenci</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2011</year>
          .
          <article-title>Composing and updating verb argument expectations: A distributional semantic model</article-title>
          .
          <source>In Proc 2nd Workshop on Cognitive Modeling and Computational Linguistics</source>
          ,
          <fpage>58</fpage>
          -
          <lpage>66</lpage>
          . ACL.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <surname>Levesque</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Davis</surname>
          </string-name>
          , E.; and
          <string-name>
            <surname>Morgenstern</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <year>2012</year>
          .
          <article-title>The Winograd Schema Challenge</article-title>
          .
          <source>In Thirteenth International Conference on the Principles of Knowledge Representation and Reasoning.</source>
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <string-name>
            <surname>Levesque</surname>
            ,
            <given-names>H. J.</given-names>
          </string-name>
          <year>2014</year>
          .
          <article-title>On our best behaviour</article-title>
          .
          <source>Artificial Intelligence</source>
          <volume>212</volume>
          :
          <fpage>27</fpage>
          -
          <lpage>35</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <string-name>
            <surname>Levesque</surname>
            ,
            <given-names>H. J.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>Common sense, the Turing test, and the quest for real AI</article-title>
          . Cambridge, MA: MIT Press.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <string-name>
            <surname>Levinson</surname>
            ,
            <given-names>S. C.</given-names>
          </string-name>
          <year>1995</year>
          .
          <article-title>Three levels of meaning. In Grammar and meaning: Essays in honour of Sir John Lyons</article-title>
          . Cambridge University Press.
          <fpage>90</fpage>
          -
          <lpage>115</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Jiangb</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Linga</surname>
            ,
            <given-names>Z.-H.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Zhuc</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Weid</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ; and Hua,
          <string-name>
            <surname>Y.</surname>
          </string-name>
          <year>2016</year>
          .
          <article-title>Commonsense Knowledge Enhanced Embeddings for Solving Pronoun Disambiguation Problems in Winograd Schema Challenge</article-title>
          .
          <source>arXiv preprint arXiv:1611</source>
          .
          <fpage>04146</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          <string-name>
            <surname>McCarthy</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>1993</year>
          .
          <article-title>Notes on formalizing context</article-title>
          .
          <source>In Proceedings of the 13th international joint conference on Artifical intelligenceVolume 1</source>
          ,
          <fpage>555</fpage>
          -
          <lpage>560</lpage>
          . Morgan Kaufmann Publishers Inc.
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          <string-name>
            <surname>McCarthy</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>1998</year>
          .
          <article-title>Elaboration tolerance</article-title>
          .
          <source>In Common Sense</source>
          , volume
          <volume>98</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Yih</surname>
            , W.-t.; and Zweig,
            <given-names>G.</given-names>
          </string-name>
          <year>2013</year>
          .
          <article-title>Linguistic regularities in continuous space word representations</article-title>
          .
          <source>In NAACL HLT</source>
          , volume
          <volume>13</volume>
          ,
          <fpage>746</fpage>
          -
          <lpage>751</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          <string-name>
            <surname>Miller</surname>
            ,
            <given-names>G. A.</given-names>
          </string-name>
          <year>1995</year>
          .
          <article-title>WordNet: a lexical database for English</article-title>
          .
          <source>Communications of the ACM</source>
          <volume>38</volume>
          (
          <issue>11</issue>
          ):
          <fpage>39</fpage>
          -
          <lpage>41</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          <string-name>
            <surname>Mitkov</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <year>2014</year>
          .
          <article-title>Anaphora resolution</article-title>
          . Routledge.
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          <string-name>
            <surname>Morgenstern</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Ortiz</surname>
            Jr,
            <given-names>C. L.</given-names>
          </string-name>
          <year>2015</year>
          .
          <article-title>The Winograd Schema Challenge: Evaluating Progress in Commonsense Reasoning</article-title>
          . In AAAI,
          <fpage>4024</fpage>
          -
          <lpage>4026</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          <string-name>
            <surname>Ng</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <year>2017</year>
          .
          <article-title>Machine Learning for Entity Coreference Resolution: A Retrospective Look at Two Decades of Research</article-title>
          . In AAAI,
          <fpage>4877</fpage>
          -
          <lpage>4884</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          <string-name>
            <surname>Peng</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Khashabi</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ; and
          <string-name>
            <surname>Roth</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <year>2015</year>
          .
          <article-title>Solving hard coreference problems</article-title>
          .
          <source>In Proceedings of NAACL</source>
          ,
          <fpage>809</fpage>
          -
          <lpage>819</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          <string-name>
            <surname>Pustejovsky</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>1991</year>
          .
          <article-title>The Generative Lexicon</article-title>
          .
          <source>Computational linguistics 17</source>
          <volume>(4)</volume>
          :
          <fpage>409</fpage>
          -
          <lpage>441</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          <string-name>
            <surname>Rahman</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Ng</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <year>2012</year>
          .
          <article-title>Resolving complex cases of definite pronouns: the winograd schema challenge</article-title>
          .
          <source>In Proceedings of the 2012 Joint Conference on EMNLP and CoNLL</source>
          ,
          <fpage>777</fpage>
          -
          <lpage>789</lpage>
          . ACL.
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          <string-name>
            <surname>Recanati</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <year>2004</year>
          .
          <article-title>Pragmatics and Semantics</article-title>
          .
          <source>In Handbook of Pragmatics. Oxford: Blackwell</source>
          .
          <fpage>442</fpage>
          -
          <lpage>462</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          <string-name>
            <surname>Rosch</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Mervis</surname>
            ,
            <given-names>C. B.</given-names>
          </string-name>
          <year>1975</year>
          .
          <article-title>Family resemblances: Studies in the internal structure of categories</article-title>
          .
          <source>Cognitive Psychology</source>
          <volume>7</volume>
          (
          <issue>4</issue>
          ):
          <fpage>573</fpage>
          -
          <lpage>605</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          <string-name>
            <surname>Rosch</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <year>1978</year>
          .
          <article-title>Principles of categorization</article-title>
          . In Rosch, E., and
          <string-name>
            <surname>Lloyd</surname>
          </string-name>
          , B. B., eds.,
          <source>Cognition and categorization</source>
          , volume
          <volume>1</volume>
          . Hillsdale, NJ: Lawrence Erlbaum Associates.
          <fpage>27</fpage>
          -
          <lpage>78</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          <string-name>
            <surname>Schu</surname>
          </string-name>
          ¨ller, P., and
          <string-name>
            <surname>Kazmi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <year>2015</year>
          .
          <article-title>Using Semantic Web Resources for Solving Winograd Schemas: Sculptures, Shelves, Envy, and Success</article-title>
          . In
          <source>SEMANTiCS (Posters &amp; Demos)</source>
          ,
          <fpage>22</fpage>
          -
          <lpage>25</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          <string-name>
            <surname>Schu</surname>
          </string-name>
          ¨ller, P.
          <year>2014</year>
          .
          <article-title>Tackling Winograd Schemas by formalizing relevance theory in knowledge graphs</article-title>
          .
          <source>In Fourteenth International Conference on the Principles of Knowledge Representation and Reasoning.</source>
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          <string-name>
            <surname>Sharma</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Vo</surname>
            ,
            <given-names>N. H.</given-names>
          </string-name>
          ; Aditya,
          <string-name>
            <given-names>S.</given-names>
            ; and
            <surname>Baral</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <year>2015</year>
          .
          <article-title>Towards Addressing the Winograd Schema Challenge-Building and Using a Semantic Parser and a Knowledge Hunting Module</article-title>
          .
          <source>In IJCAI</source>
          ,
          <fpage>1319</fpage>
          -
          <lpage>1325</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          <string-name>
            <surname>Sharma</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <year>2014</year>
          .
          <article-title>Solving Winograd schema challenge: Using semantic parsing, automatic knowledge acquisition and logical reasoning</article-title>
          .
          <source>Ph.D. Dissertation</source>
          , Arizona State University.
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          <string-name>
            <surname>Speer</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Havasi</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <year>2012</year>
          .
          <article-title>Representing General Relational Knowledge in ConceptNet 5</article-title>
          .
          <string-name>
            <surname>In</surname>
            <given-names>LREC</given-names>
          </string-name>
          ,
          <fpage>3679</fpage>
          -
          <lpage>3686</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          <string-name>
            <surname>Sperber</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , and Wilson,
          <string-name>
            <surname>D.</surname>
          </string-name>
          <year>2004</year>
          .
          <article-title>Relevance theory</article-title>
          .
          <source>In Handbook of Pragmatics. Oxford: Blackwell</source>
          .
          <fpage>607</fpage>
          -
          <lpage>632</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref43">
        <mixed-citation>
          <string-name>
            <surname>Turing</surname>
            ,
            <given-names>A. M.</given-names>
          </string-name>
          <year>1950</year>
          .
          <article-title>Computing machinery and intelligence</article-title>
          .
          <source>Mind</source>
          <volume>59</volume>
          (
          <issue>236</issue>
          ):
          <fpage>433</fpage>
          -
          <lpage>460</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref44">
        <mixed-citation>
          <string-name>
            <surname>Verheyen</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Ameel</surname>
            , E.; and Storms,
            <given-names>G.</given-names>
          </string-name>
          <year>2007</year>
          .
          <article-title>Determining the dimensionality in spatial representations of semantic concepts</article-title>
          .
          <source>Behavior Research Methods</source>
          <volume>39</volume>
          (
          <issue>3</issue>
          ):
          <fpage>427</fpage>
          -
          <lpage>438</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref45">
        <mixed-citation>
          <string-name>
            <surname>Zadeh</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <year>1965</year>
          .
          <article-title>Fuzzy sets</article-title>
          .
          <source>Information and Control</source>
          <volume>8</volume>
          (
          <issue>3</issue>
          ):
          <fpage>338</fpage>
          -
          <lpage>353</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>