<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Phase-based Minimalist Parsing and complexity in non-local dependencies</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Cristiano Chesi NETS - IUSS P.zza Vittoria</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pavia (Italy)</string-name>
        </contrib>
      </contrib-group>
      <abstract>
        <p>English. A cognitively plausible parsing algorithm should perform like the human parser in critical contexts. Here I propose an adaptation of Earley's parsing algorithm, suitable for Phase-based Minimalist Grammars (PMG, Chesi 2012), that is able to predict complexity effects in performance. Focusing on self-paced reading experiments of object clefts sentences (Warren &amp; Gibson 2005) I will associate to parsing a complexity metric based on cued features to be retrieved at the verb segment (Feature Retrieval &amp; Encoding Cost, FREC). FREC is crucially based on the usage of memory predicted by the discussed parsing algorithm and it correctly fits with the reading time revealed.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        Italian. Un algoritmo di parsing
cognitivamente plausibile dovrebbe avere una
performance paragonabile a quella
umana in contesti critici. In questo lavoro
propongo un adattamento dell’algoritmo
di Earley che utilizza Grammatiche
Minimaliste basate sul concetto di Fase
        <xref ref-type="bibr" rid="ref2">(PMG,
Chesi 2012)</xref>
        . Associata all’algoritmo,
verrà discussa una funzione di costo
(Feature Retrieval &amp; Encoding Cost, FREC)
capace di misurare la difficoltà relativa al
recupero dei referenti coinvolti in
dipendenze a distanza. La funzione si basa sui
tratti morfosintattici archiviati nel
memory buffer utilizzato dal parser.
Concentrandosi sulle strutture scisse ad
estrazione dell’oggetto, si mostrerà come il
      </p>
    </sec>
    <sec id="sec-2">
      <title>FREC risulti predittivo dei dati sperimentali ricavati da studi classici di lettura autoregolata (Warren &amp; Gibson 2005).</title>
      <p>1</p>
      <sec id="sec-2-1">
        <title>Introduction</title>
        <p>
          The last twenty years of formal linguistic research
have been deeply influenced by Chomsky’s
minimalist intuitions
          <xref ref-type="bibr" rid="ref4 ref6">(Chomsky 1995, 2013)</xref>
          . In a
nutshell, the core Minimalist proposal is to reduce
phrase structure formation to the recursive
application of a binary, bottom-up, structure-building
operation dubbed Merge. Merge creates
hierarchical structures by combining two lexical items
(1.a), one lexical item and an already built (by
previous application of Merge operations) phrase
(1.b) or two already built phrases (1.c).
        </p>
        <p>(1) a.</p>
        <p>x
y
b.
x</p>
        <p>YP</p>
        <p>c.</p>
        <p>XP</p>
        <p>
          YP
Phrases are not linearly ordered by Merge. Only
when they are spelled-out (i.e. sent to the
SensoryMotor interface, aka Phonetic Form, PF),
linearization is required: assuming that x and y are
terminal nodes (i.e. words), either &lt;x, y&gt; or &lt;y, x&gt;
can both be proper linearizations of (1.a).
Hierarchical structure (and linearization) is also
determined by another structure building operation:
Move
          <xref ref-type="bibr" rid="ref4">(or Internal Merge, Chomsky 1995)</xref>
          ; Move
re-arranges phrases in the structure by re-merging
an item (already merged in the structure) to the
edge of the current, top-most, phrase: for instance
[XP [YP [ZP]]] can lead to [ZP [XP [YP (ZP)]] if
XP (the probe) has a feature triggering movement
(e.g. +f) and ZP (the goal) has the relevant feature
qualifying it as a plausible target for movement
(e.g. -f). At the end, the element displaced (ZP)
will occupy the edge of the structure. When the
items within an already built phrase, for instance
XP, are delivered to PF, they get properly
linearized according to their hierarchical structure
          <xref ref-type="bibr" rid="ref18 ref20">(e.g.
Linear Correspondence Axiom, Kayne 1994)</xref>
          ,
intrinsic phonetic properties (e.g. cliticization), as
well as economy conditions (e.g. an items should
not be pronounced twice). Such a (cyclic)
spellout happens at phases: XP will be delivered to PF
only if it qualifies as a phase
          <xref ref-type="bibr" rid="ref6">(Chomsky 2013)</xref>
          . In
this sense, a phase should be a constituent/phrase
with some degree of completeness with respect to
semantic interpretation (Logic Form, aka LF).
Most minimalist linguists agree on the fact that a
full-fledged sentence (aka Complementizer
Phrase, CP) is a phase, the highest argumental
shell of a predicate qualifies as a phase (aka
littlev Phrase, vP) and also a full argument is a phase
(aka Determiner Phrase, DP). Such a simple (and
computationally appealing) model has been fully
formalized
          <xref ref-type="bibr" rid="ref25 ref7">(Stabler 1997, Collins &amp; Stabler 2016)</xref>
          and some parsing algorithm that implements main
minimalist insights has been discussed in
literature
          <xref ref-type="bibr" rid="ref15 ref17 ref2">(e.g. Harkema 2001, Chesi 2012 a.o.)</xref>
          .
        </p>
        <p>
          In these pages, I will present some of the
advantages of retaining such a simplified
computational approach to syntactic derivation. Crucially,
I will try to overcome some clear disadvantages in
assuming the just presented standard, bottom-up,
structure building operations, while obtaining, at
the same time, a better empirical fit: on the one
hand, I will avoid any non-efficient
deductiveparsing perspective (that is a consequence of the
assumed bottom-up nature of the Merge and
Move operations); on the other, I will promote a
more transparent relation between formal
competence, parsing and psycholinguistic performance
by presenting a simple adaptation of Earley’s
Top-Down parsing algorithm
          <xref ref-type="bibr" rid="ref8">(Earley 1970)</xref>
          and a
complexity metric that refers directly to parsing
memory usage: this metric will be able to account
for complexity in retrieving the correct item while
processing specific non-local dependencies. By
“non-local” dependencies I refer to those relations
involving movement, namely constructions where
the very same item occurs in two distinct,
non-adjacent, positions: for instance, wh-dependencies in
English require the wh- item (who, in (1)) to be
interpreted both in a the left peripheral (focalized)
position
          <xref ref-type="bibr" rid="ref24">(the Criterial position, in the sense of
Rizzi 2007)</xref>
          and in the thematic lower position
(right next to the verb meet in (1))1:
(1) Who1 do you think Mary will meet _1?
with wh-questions share a similar non-local
dependency formation:
(2) a. It is [DP1 the banker|John|me] that
        </p>
        <p>
          [DP2 the lawyer|Dan|you] will meet _DP1
In short, the head of the dependency (DP1) should
be interpreted both as a focalized item and as the
direct object (this is where the name of the
construction “object cleft” comes from) of the
embedded verb. The difficulty of parsing this structure
has been deeply discussed in literature
          <xref ref-type="bibr" rid="ref14">(Gordon et
al. 2004)</xref>
          . What is considered a crucial factor is the
role of the similarity between DP1 and DP2
          <xref ref-type="bibr" rid="ref1 ref6">(the
subject of the cleft, Belletti and Rizzi 2013, §2)</xref>
          .
To capture this fact, I will re-adapt Earley’s
algorithm (§3.1) to operate on a specific version of
Minimalist Grammar (§3). This would allow us to
subsume the similarity effect by predicting
reading differences as revealed in self-paced reading
experiments
          <xref ref-type="bibr" rid="ref27">(e.g. Warren &amp; Gibson 2005, §4)</xref>
          .
2
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Parsing with Minimalist Grammars</title>
        <p>
          Since Merge and Move strictly operate “from
bottom to top”, we expect sentence structure in (2) to
be built in 9 steps (and 5 phases: ph1, ph2 …):
1. [ph1 the banker]
2. [ph3 meet [ph1 …]]]
3. [ph3 will [meet [ph1 …]]]
4. [ph2 the lawyer] (independently built)
5. [ph3 [ph2 …] will [meet [ph1 …]]]
6. [ph4 that [ph3 [ph2 …] will [meet [ph1 …]]]]
7. [[ph1 …] [ph4 that [ph3 [ph2 …] will [meet (ph1 …)]]]]
(ph1 moves to ph4 edge)
8. [ph5 is [[ph1 …] [ph4 that [ph3 [ph2 …] will [meet
[ph1 …]]]]]
9. [ph5 it [is [[ph1 …] [ph4 that [ph3 [ph2 …] will [meet
[ph1 …]]]]]
With the exception of step 4, all other steps must
be strictly ordered. As a consequence, moving the
direct object in the relevant position would force
the linearization to place ph2 first at the edge of
ph3, then at the edge of ph4. This is how
Minimalism derives the relevant non-local dependencies in
(2). Obviously this is not transparent at all with
respect to parsing
          <xref ref-type="bibr" rid="ref10">(e.g. Fong 2011)</xref>
          , where the
processing order is expected to be completely
reversed:
        </p>
        <p>
          The critical derivation I will discuss in this
paper is that of object clefts
          <xref ref-type="bibr" rid="ref13">(Gordon et al. 2001)</xref>
          that
1. [ph5 ] is initiated
2. [ph1 ] is fully processed while [ph5 ] is still open
1 Coreference in non-local dependencies will be
indicated by the same subscript placed both on the
“displaced” item and on the thematic position (the
nonpronounced item in the thematic position is indicated
with a co-indexed underscore)
3. [ph4 ] is initiated (a Relative Clause)
4. [ph3 ] is initiated as well (Verbal Phrase)
5. [ph2 …] is fully processed while [ph5 ], [ph4 ] and [ph3 ] are
open
6. [ph1 ] finally receives a thematic role, hence [ph5 ], [ph4 ]
and [ph3 ] can be closed.
        </p>
        <p>
          Unless we deeply revise Minimalist Grammars
          <xref ref-type="bibr" rid="ref22 ref9">(both with respect to movement, Fong 2005, and
to thematic role assignment, Niyogi &amp; Berwick
2005)</xref>
          , we are left with an asymmetry that can not
be explained simply in terms of structure building
operations as discussed in the next section.
2.1
        </p>
        <sec id="sec-2-2-1">
          <title>The “similarity” problem</title>
          <p>
            <xref ref-type="bibr" rid="ref27">Warren &amp; Gibson (2005)</xref>
            show that in clefts
constructions like the one discussed in (2), the
variation of the two DPs [ph1 ] and [ph2 ] produces
differences in reading time at the verb segment in
self-paced reading experiments with the full-DP
matching condition ([ph1 the barber] that [ph2 the
banker] praised …) and proper nouns matching
condition ([ph1 John] that [ph2 Dan] praised …)
ranking higher in terms of difficulty (greatest slow
down at verb segment), while pronouns ([ph1 you]
that [ph2 we] praised …) are easier (fastest reading
time). No CFG-based parsing algorithm (in fact,
no classic algorithm implements the non-local
dependencies in (2) as presented in §2) or
Minimalist deductive parsing
            <xref ref-type="bibr" rid="ref21">(parsing strategies exploit
the weak equivalence of MGs with multiple
Context Free Grammars, Michaelis 1998)</xref>
            have a
chance to compare these cases.
3
          </p>
        </sec>
      </sec>
      <sec id="sec-2-3">
        <title>A processing-friendly proposal</title>
        <p>
          Phase-based Minimalist Grammars
          <xref ref-type="bibr" rid="ref2">(PMGs, Chesi
2012)</xref>
          suitable for parsing of sentences like the
ones in (2) can be formalized as follows:
(3) PMG able to parse cleft sentences
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Lexicon</title>
      <p>[[+D +Sg Johni] [N _i]], [[+D +Sg Dani] [N _i]], [N +Sg banker],
[N +Sg lawyer], [+D the], [+D +P1 +Pl +case_acc me [N Ø]],
[+D +P2 +Sg +case_nom you [N Ø]], [+T will], [+T that],
[=[DP (+case_nom)] =[DP (+case_acc)] V meet], [+exp it], [=rCP BE is]</p>
    </sec>
    <sec id="sec-4">
      <title>Phases</title>
      <p>DP
Cleft
rCP
→
→
→
[DP ([+F Ø]/[+S Ø]) +D N]
[CP +Exp BE]
[CP +F +FIN (+S) +T V]</p>
    </sec>
    <sec id="sec-5">
      <title>Operations</title>
      <p>
        Merge = ([phH +f (+fn) (H)], [+f L]) = [phH [+f L (+fn) (H)]]
Phase Projection = [phH =phX H] = [phH =phX H [phX ]]
Move = if expected [phX +f X] and found
[phX [phY +f +g Y] X] → MEM([phY +g &lt;Y&gt;)])
As in MGs
        <xref ref-type="bibr" rid="ref25">(Stabler 1997)</xref>
        , the Lexicon is a finite
set of lexical items storing phonetic, semantic
(here ignored) and syntactic features (functional
+F, selectional =S, categorial C); an item bearing
a selection feature, e.g. [=XP A], requires an XP
ph(r)ase right afterward: [=XP A [XP ]] (once
features are projected in the structure, i.e. [XP ], the
selection features are deleted, i.e. =XP); functional
features, e.g. +X express a functional
specification like determiner +D, tense +T or topic +S
(when placed under brackets, e.g. (+f), functional
features are optional; Ø indicates phonetically null
items).
      </p>
      <p>Merge simply unifies the expected structure built
so far with a new incoming item, if and only if,
this item bears (at least) the first relevant feature
expected (Merge operation is greedy: an item
bearing more features in the correct expected
order will lexicalize them all):
1. Merge([+X +Y +Z W ], [+X +Y A])=[[+X +Y A] +Z W ]
2. Merge([[+X +Y A] +Z W ], [+z B])=[[+X +Y A][+z B] W ]
3. Merge([[+X +Y A][+z B] W ], [w C])=[[+X +Y A][+z B] [w C]]
Move uses a Last-In-First-Out (LIFO) memory
buffer (M) to create non-local dependencies: M is
used to store unexpected bundles of features
merged in the derivation (below, underlined
features, e.g. [+W U], are the unexcepted ones
triggering Move):
1'. Merge([+X +Y +Z W ], [+X +W U A]) = [[+X +W U A] +Z W ]
2'. Move([+X +W U A]) = M[+W U &lt;A&gt;]
Items in the memory buffer M will be re-merged
in the structure, before any other item taken from
the lexicon, as soon as a coherent selection is
introduced by another merged item:
3'. Merge([ … [w =[+W U] C [+W U ]]], M[+W U &lt;A&gt;]) =
[ … [w =[+W U] C [+W U &lt;A&gt;)]]], M[ empty ]
Notice that phonetic features (items under angled
brackets, i.e. [&lt;A&gt;]) are not re-merged in the
structure (that is, they are not expected to be found in
the input) since they are already been
pronounced/parsed in the higher position. When the
M(emory) buffer is empty and no more selection
features must be expanded, the procedure ends.
3.1</p>
      <sec id="sec-5-1">
        <title>Parsing cleft structures with PMGs</title>
        <p>The parsing algorithm using the minimalist
grammar described in (3) implements an Earley-like
procedure composed of three sub-routines:</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Ph(ase)P(rojection) (Earley Prediction proce</title>
      <p>dure): the most prominent (i.e. first/left most)
select feature is expanded (the sentence parsing
starts with a default PhP using one of the phases in
grammar (3));
Merge (Earley Scanning procedure): if Memory is
empty, the first available feature F in the expected
phase is searched in the input/lexicon and possible
items will be retrieved2 (search(F) = [F lex1], [F
lex2] … [F lexn]) then unified with the expected
structure (e.g. Merge([F … X], [F lex1]) = [[F lex1]…
x]); items stored in Memory are checked before the
sentence input for Merge;
Move: if more features than the one expected are
introduced, those features are clustered and moved
in the LIFO Memory buffer:</p>
      <p>M[[slot 1][slot 2] … [slot n]].</p>
      <p>Given the recursive, cyclic, application of the
three subroutines above, this is the sequence of
steps needed for parsing a cleft sentence like (2):
1. Default PhP (in this case: Cleft): [CP +Exp BE]
2. Search(+Exp): M[ empty ], Lex[[+exp it]]
3. Merge([CP +Exp BE], [+exp it]) = [CP [+exp it] BE]
4. Search(BE): M[ empty ], Lex[[BE is]]
5. Merge([CP [+exp it] BE], [=rCP BE is]) =</p>
      <p>[CP [+exp it] [=rCP BE is]]
6. PhP([CP [+exp it] [=rCP BE is]) =</p>
      <p>[CP [+exp it] [=rCP BE is [CP +F +FIN +S +T V]]
7. Search(+F): M[ empty ], Lex[[DP [+F Ø] +D N]]
8. Merge([…[CP +F +FIN +S +T V]], [DP [+F Ø] +D N]) =
[CP [DP [+F Ø] +D N] +FIN +S +T V]]
9. Search(+D): M[ empty ], Lex[[+D the]]
10. Merge([DP [+F Ø] +D N], [+D the]) =</p>
      <p>[CP [DP [+F Ø] [+D the] N] +FIN +S +T V]]
11. Search(N): M[ empty ], Lex[[N banker]]
12. Merge([DP [+F Ø] [+D the] N], [N banker]) =</p>
      <p>[CP [DP [+F Ø] [+D the] [N banker]] +FIN +S +T V]]
13. Move([DP [+F Ø] [+D the] N], [N banker]) =</p>
      <p>M[[DP +D N &lt;the banker&gt;]]
(Move is triggered because at step 8 +D N were
unexpected; only after full lexicalization [DP [+F Ø]
+D N] is stored in M, namely at step 13)
14. Search(+FIN): M[[DP +D N &lt;the banker&gt;]], Lex[[+FIN that]]
15. Merge([CP [DP [+F Ø] [+D the] [N banker]] +FIN +S +T V]],
[+FIN that]) = [CP [DP [+F Ø] [+D the] [N banker]] [+FIN that]
+S +T V]]
16. Search(+S): M[[DP +D N &lt;the banker&gt;]],</p>
      <p>Lex[[DP [+S Ø] +D N]]
17. Merge([CP [DP [+F Ø] [+D the] [N banker]] [+FIN that] +S +T
V]], [DP [+S Ø] +D N]) = [CP [DP [+F Ø] [+D the] [N banker]]
[+FIN that] [DP [+S Ø] +D N] +T V]]
18. (repeat 9-13 mutatis mutandis)
19. Search(+T): M[[DP +D N (the lawyer)],[DP +D N (the
banker)]], Lex[+T will]
2 For reason of space, I will not discuss here neither
lexical and syntactic ambiguity nor reanalysis (i.e.
recovery from wrong expectations); the proposed
algorithm here is meant to be a Top-Down complete
procedure, that is, all the possible ambiguities will be
20. Merge([CP [DP [+F Ø] [+D the] [N banker]] [+FIN that] [DP
[+S Ø] [+D the] [N lawyer]] +T V]], [+T will]) =
([CP [DP [+F Ø] [+D the] [N banker]] [+FIN that] [DP [+S Ø]
[+D the] [N lawyer]] [+T will] V]]
21. Search(V): M[[DP +D N (the lawyer)],[DP +D N (the
banker)]], Lex[=DP =DP V meet]
22. Merge([CP [DP [+F Ø] [+D the] [N banker]] [+FIN that] [DP
[+S Ø] [+D the] [N lawyer]] [+T will] V]], [=DP =DP V meet])
=
[CP [DP [+F Ø] [+D the] [N banker]] [+FIN that] [DP [+S Ø]
[+D the] [N lawyer]] [+T will] [=DP =DP V meet]]
23. PhP([CP … [=DP =DP V meet]]) = [CP … [=DP =DP V meet
[DP +D N]]]
24. Merge ([CP … [=DP =DP V meet [DP +D N]]], M[[DP +D N
(the lawyer)]] = ([CP … [=DP =DP V meet [DP +D N (the
lawyer)]]]
25. PhP([CP … [=DP =DP V meet]]) = [CP … [=DP =DP V meet
[DP +D N (the lawyer)] [DP +D N]]]
26. Merge ([CP … [=DP =DP V meet … [DP +D N]]], M[[DP +D N
(the banker)]] = ([CP … [=DP =DP V meet … [DP +D N (the
banker)]]]
According to the lexicon and the phase
expectations, step 10 and 19 could have found in the input
[+D N John], [+D N Dan], [+D +P1 +Pl +case_acc me [N Ø]]
or [+D +P2 +Sg +case_nom you [N Ø]], capturing all
possible combinations of definite descriptions,
correct pronominal DPs and proper nouns. Exactly all
the possibilities we want to test.</p>
      <sec id="sec-6-1">
        <title>4 Explaining the “similarity” problem in terms of cue-based feature retrieval</title>
        <p>
          According to
          <xref ref-type="bibr" rid="ref27">Warren &amp; Gibson (2005)</xref>
          revealed
reading times
          <xref ref-type="bibr" rid="ref14">(see also Gordon et al. 2004 for very
similar results)</xref>
          we can roughly rank on a difficulty
scale all the (3x3) tested conditions (D = definite
condition, e.g. “the banker”, N = nominal
condition, e.g. “Dan”, P = pronoun condition, e.g. “we”;
for instance D-D stands for “it is the banker that
the lawyer will meet…”, vs D-P condition “it is
the banker that we will meet…”):
(4) D-D ≥ N-D ≈ N-N ≈ P-D
&gt; D-N ≥ P-N &gt; D-P ≥ N-P ≈ P-P
 
Building on
          <xref ref-type="bibr" rid="ref12">Gillund &amp; Shiffrin (1984)</xref>
          Search of
Associative Memory (SAM) model, and
assuming a cue-based retrieval mechanism for items in
memory
          <xref ref-type="bibr" rid="ref26">(Van Dyke &amp; McElree 2006)</xref>
          , we can
define a complexity (C) function associated to the
features to be retrieved from M (Feature Retrieval
taken into consideration and stored in the parsing
“chart” as in the classic Earley’s parser. For ranking
of alternatives see
          <xref ref-type="bibr" rid="ref15">Hale (2001)</xref>
          .
        </p>
        <p>
          Cost, FRC,
          <xref ref-type="bibr" rid="ref3">Chesi 2016</xref>
          ) for each item to be
remerged after the phase projection at verb (V):
(5) CFRC(V) = ∏
In the formula above, m is number of items stored
in memory at retrieval, nF is the number of
features characterizing the argument to be retrieved
that are non-distinct in memory (i.e. also present
in other objects in memory), dF is number of
distinct cued features (e.g. case features explicitly
probed by the verb selection). CFRC will express
the cost, in numerical terms, that should fit with
the revealed reading time (i.e. higher differences
in reading times, higher differences in CFRC).
According to the lexicon in (3), the cost for
retrieving the correct items in the D-D condition, for
instance, is calculated as follows:
1. [=[DP (+case_nom)] =DP(+case_acc) V meet] will trigger
retrieval of the first item (the last inserted one
in the buffer) which is (step 24) the DP
[+D +Sg N (the lawyer)]
2. No cued-features are present (the verb
selection only asks for an optional nominative
case) and the 3 features to be retrieved are in
fact shared with the other item in memory
([+D +Sg N (the banker)])
3.
        </p>
        <p>
          Hence: CFRC =
x
Notice that retrieving the object when the subject
has been removed from memory has a minimal
cost since no confounding features are present
anymore in memory. As for the other relevant
conditions: N-N, as in D-D condition share the same
features hence we expect them to have similar cost
except for the fact that N feature is not fully
lexicalized, but it is a trace of an N-to-D movement
          <xref ref-type="bibr" rid="ref20">(Longobardi 1994)</xref>
          . Counting this as 0.5 (further
investigation is needed to correctly assign a cost
to an emptied lexical position), we obtain 12,25.
Same complexity for N-D condition (since the [N]
lexical feature in D is compared to the trace [N _i]
feature of N counting 0.5). While we would
expect slightly smaller cost with the P-D condition
(P does have a [N Ø] empty feature), that is 9, we
will correctly predict simpler complexity for
retrieving pronouns at the subject position, since
they are always bearing person features (which
are distinct from default 3rd person of D and N)
and they are marked for case (which is cued by the
verb, producing the minimal cost in the P-P
condition (CFRC= 1) and similar costs in the D-P and
N-P conditions (both CFRC=4). Predictions can be
further differentiated by adding a cost for
encoding the features in the structure (eF) which is (to
keep the calculation as simple as possible)
proportional to the number of lexical features to be
encoded once an item is retrieved from memory (the
numerator of the CFRC cost function becomes:
1 ). This corresponds to an
increase of +1 for D and +0,5 for N at retrieval. The
new CFREC(V) in the different conditions becomes:
CFREC(V) D-D =
CFREC(V) N-D =
CFREC(V) N-N = ,
CFREC(V) P-D =
CFREC(V) D-N = ,
CFREC(V) P-N = ,
CFREC(V) D-P =
CFREC(V) N-P =
CFREC(V) P-P =
        </p>
        <p>
          x ,
x
x ,
x
x
x
x
x ,
x
Though in some cases FREC predicts slightly
larger differences (e.g. D-D vs N-D/N-N
condition), it correctly ranks all conditions revealed by
the discussed experiment, and it is coherent with
specific predictions (e.g. related to feature
matching) discussed in literature
          <xref ref-type="bibr" rid="ref1">(Belletti &amp; Rizzi 2013)</xref>
          .
5
        </p>
        <sec id="sec-6-1-1">
          <title>Conclusion</title>
          <p>
            In this paper I presented an adaptation of Earley’s
Top-Down parsing algorithm to be used with a
simple implementation of a Minimalist Grammar
(PMG). The advantages of this approach are both
in terms of cognitive plausibility and
parsing/performance transparency. From the cognitive
plausibility perspective, I showed how a re-orientation
of the minimalist structure building operations
Merge and Move is sufficient to include such
operations directly within a parsing procedure. This
is a step toward the “Parser Is the Grammar”
(PIG) default hypothesis (Phillips 2006) and a
welcome simplification of the linguistic
competence description: such a grammar description (i.e.
our linguistic competence) is shared both in
production (generation) and in comprehension
(parsing); this seems trivial from a cognitive
perspective (we have a unique Broca’s area activated in
syntactic processing both in parsing and in
generation), but it is far from trivial in computational
terms. On the other hand, from the
parsing/performance transparency perspective, I presented a
complexity metric (FREC), based on cued
features stored in memory which better characterize
performance in object clefts constructions
compared to alternative models: for instance the
Depencency Locality Theory (DLT) based on
accessibility hierarchy
            <xref ref-type="bibr" rid="ref11">(Gibson 2000)</xref>
            is unable to
predict high complexity in N-N condition
comparable to N-D or D-D condition, since N should be
uniformly more accessible than D, contrary to the
facts. The proposed model, obviously should be
extended in many respects to capture other critical
phenomena
            <xref ref-type="bibr" rid="ref19">(see Lewis &amp; Vasishth 2005)</xref>
            but the
first results on specific well-studied constructions,
like object clefts, seem very promising.
          </p>
        </sec>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Belletti</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Rizzi</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <year>2013</year>
          .
          <article-title>Intervention in grammar and processing. From grammar to meaning: The spontaneous logicality of language</article-title>
          ,
          <fpage>293</fpage>
          -
          <lpage>311</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Chesi</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <year>2012</year>
          .
          <article-title>Competence and Computation: toward a processing friendly minimalist Grammar</article-title>
          . Padova: Unipress.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>Chesi</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <year>2016</year>
          .
          <article-title>Il processamento in tempo reale delle frasi complesse</article-title>
          .
          <source>EM Ponti, M. Budassi (eds.)</source>
          ,
          <fpage>21</fpage>
          -
          <lpage>38</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Chomsky</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <year>1995</year>
          .
          <article-title>The Minimalist Program (Current Studies in Linguistics 28)</article-title>
          . Cambridge (MA): MIT Press.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Chomsky</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <year>2008</year>
          .
          <article-title>On phases</article-title>
          . In Robert Freidin,
          <string-name>
            <given-names>Carlos P.</given-names>
            <surname>Otero</surname>
          </string-name>
          , and Maria Luisa Zubizarreta (eds.).
          <source>Foundational issues in linguistic theory</source>
          ,
          <volume>133</volume>
          -
          <fpage>166</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Chomsky</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <year>2013</year>
          .
          <article-title>Problems of projection</article-title>
          .
          <source>Lingua</source>
          ,
          <volume>130</volume>
          ,
          <fpage>33</fpage>
          -
          <lpage>49</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Collins</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Stabler</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <year>2016</year>
          .
          <article-title>A formalization of minimalist syntax</article-title>
          .
          <source>Syntax</source>
          ,
          <volume>19</volume>
          (
          <issue>1</issue>
          ),
          <fpage>43</fpage>
          -
          <lpage>78</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Earley</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>1970</year>
          .
          <article-title>An efficient context-free parsing algorithm</article-title>
          .
          <source>Communications of the Association for Computing Machinery</source>
          ,
          <volume>13</volume>
          (
          <issue>2</issue>
          ), February.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Fong</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <year>2005</year>
          .
          <article-title>Computation with probes and goals</article-title>
          . In Di Sciullo,
          <string-name>
            <surname>A. M.</surname>
          </string-name>
          and
          <string-name>
            <surname>R</surname>
          </string-name>
          . Delmonte (eds.).
          <article-title>UG and external systems: Language, brain and computation</article-title>
          .
          <volume>75</volume>
          ,
          <fpage>311</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>Fong</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <year>2011</year>
          .
          <article-title>Minimalist parsing: Simplicity and feature unification</article-title>
          .
          <source>In Proceedings of Workshop on Language and Recursion</source>
          . Mons, Belgium: University of Mons. March.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>Gibson</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <year>2000</year>
          .
          <article-title>The dependency locality theory: A distance-based theory of linguistic complexity. Image, language</article-title>
          , brain,
          <fpage>95</fpage>
          -
          <lpage>126</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <surname>Gillund</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Shiffrin</surname>
            ,
            <given-names>R. M.</given-names>
          </string-name>
          <year>1984</year>
          .
          <article-title>A retrieval model for both recognition and recall</article-title>
          .
          <source>Psychological review</source>
          ,
          <volume>91</volume>
          (
          <issue>1</issue>
          ),
          <fpage>1</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <surname>Gordon</surname>
            ,
            <given-names>P. C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hendrick</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Johnson</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <year>2001</year>
          .
          <article-title>Memory interference during language processing</article-title>
          .
          <source>Journal of Experimental Psychology: Learning, Memory, and Cognition</source>
          ,
          <volume>27</volume>
          (
          <issue>6</issue>
          ),
          <fpage>1411</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <surname>Gordon</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hendrick</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , Johnson,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <year>2004</year>
          .
          <article-title>“Effects of noun phrase type on sentence complexity”</article-title>
          ,
          <source>Journal of Memory and Language</source>
          <volume>51</volume>
          ,
          <fpage>97</fpage>
          -
          <lpage>114</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <surname>Hale</surname>
            ,
            <given-names>J. T.</given-names>
          </string-name>
          <year>2001</year>
          .
          <article-title>A probabilistic Earley parser as a psycholinguistic model</article-title>
          .
          <source>In Proceedings of the second meeting of the North American Chapter of the Association for Computational Linguistics on Language technologies</source>
          ,
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <surname>Hale</surname>
            ,
            <given-names>J. T.</given-names>
          </string-name>
          <year>2011</year>
          .
          <article-title>What a rational parser would do</article-title>
          .
          <source>Cognitive Science</source>
          ,
          <volume>35</volume>
          (
          <issue>3</issue>
          ),
          <fpage>399</fpage>
          -
          <lpage>443</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <surname>Harkema</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <year>2001</year>
          .
          <article-title>Parsing minimalist languages</article-title>
          . University of California, Los Angeles.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <surname>Kayne</surname>
            ,
            <given-names>R. S.</given-names>
          </string-name>
          <year>1994</year>
          .
          <article-title>The antisymmetry of syntax</article-title>
          . Cambridge (MA), MIT Press.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <surname>Lewis</surname>
            ,
            <given-names>R. L.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Vasishth</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <year>2005</year>
          .
          <article-title>An activation‐based model of sentence processing as skilled memory retrieval</article-title>
          .
          <source>Cognitive science</source>
          ,
          <volume>29</volume>
          (
          <issue>3</issue>
          ),
          <fpage>375</fpage>
          -
          <lpage>419</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <string-name>
            <surname>Longobardi</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <year>1994</year>
          .
          <article-title>Reference and proper names: a theory of N-movement in syntax and logical form</article-title>
          .
          <source>Linguistic Inquiry</source>
          ,
          <volume>25</volume>
          ,
          <fpage>609</fpage>
          -
          <lpage>65</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <string-name>
            <surname>Michaelis</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <year>1998</year>
          .
          <article-title>Derivational Minimalism is Mildly Context-Sensitive.</article-title>
          ” In M. Moortgat (ed.),
          <source>Logical Aspects of Computational Linguistics</source>
          ,
          <source>(LACL '98). Lecture Notes in Artificial Intelligence</source>
          . Springer Verlag.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <string-name>
            <surname>Niyogi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Berwick</surname>
            ,
            <given-names>R. C.</given-names>
          </string-name>
          <year>2005</year>
          .
          <article-title>A minimalist implementation of Hale-Keyser incorporation theory</article-title>
          . In
          <string-name>
            <surname>A. M. Di</surname>
          </string-name>
          <article-title>Sciullo (ed.) UG and external systems language, brain and computation</article-title>
          , linguistik aktuell/linguistics today,
          <volume>75</volume>
          ,
          <fpage>269</fpage>
          -
          <lpage>288</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          <string-name>
            <surname>Phillips</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <year>1996</year>
          .
          <article-title>Order and structure</article-title>
          .
          <source>Doctoral dissertation</source>
          , Massachusetts Institute of Technology.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          <string-name>
            <surname>Rizzi</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <year>2007</year>
          .
          <article-title>On some properties of criterial freezing</article-title>
          .
          <source>Studies in linguistics</source>
          ,
          <volume>1</volume>
          ,
          <fpage>145</fpage>
          -
          <lpage>158</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          <string-name>
            <surname>Stabler</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <year>1997</year>
          .
          <article-title>Derivational minimalism</article-title>
          .
          <source>In International Conference on Logical Aspects of Computational Linguistics</source>
          (pp.
          <fpage>68</fpage>
          -
          <lpage>95</lpage>
          ). Springer, Berlin, Heidelberg.
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          <string-name>
            <surname>Van Dyke</surname>
            ,
            <given-names>J. A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>McElree</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <year>2006</year>
          .
          <article-title>Retrieval interference in sentence comprehension</article-title>
          .
          <source>Journal of Memory and Language</source>
          ,
          <volume>55</volume>
          (
          <issue>2</issue>
          ),
          <fpage>157</fpage>
          -
          <lpage>166</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          <string-name>
            <surname>Warren</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Gibson</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <year>2005</year>
          .
          <article-title>Effects of NP type in reading cleft sentences in English</article-title>
          .
          <source>Language and Cognitive Processes</source>
          ,
          <volume>20</volume>
          (
          <issue>6</issue>
          ),
          <fpage>751</fpage>
          -
          <lpage>767</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>