<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Visual gender cues elicit agent expectations: different mismatches in situated language comprehension.</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Michele Burigo (mburigo@cit-ec.uni-bielefeld.de)</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pia Knoeferle (knoeferl@cit-ec.uni-bielefeld.de)</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Cognitive Interaction Technology Excellence Cluster, Bielefeld University</institution>
          ,
          <addr-line>Bielefeld</addr-line>
          ,
          <country country="DE">Germany</country>
        </aff>
      </contrib-group>
      <fpage>234</fpage>
      <lpage>239</lpage>
      <abstract>
        <p>Previous research has shown that visual cues (depicted events) can have a strong effect on language comprehension and guide attention more than stereotypical thematic role knowledge ('depicted / recent event preference'). We examined to which extent this finding generalizes to another visual cue (gender from the hands of an agent) and to which extent it is modulated by picture-sentence incongruence. Participants inspected videos of hands performing an action, and then listened to non-canonical German OVS sentences while we monitored their eye gaze to the faces of two potential subjects / agents (one male and one female). In Experiment 1, the sentential verb phrase matched (vs. mismatched) the video action and in Experiment 2, the sentential subject matched (vs. mismatched) the gender of the agent's hands in the video. Additionally, both experiments manipulated gender stereotypicality congruence (i.e. whether the gender stereotypicality of the described actions matched or mismatched the gender of the hands in the video). Participants overall preferred to inspect the target agent face (i.e. the face whose gender matched that of the hands seen in the previous video), suggesting the depicted event preference observed in previous studies generalizes to visual gender cues. Stereotypicality match did not seem to modulate this gaze behavior. However, when there was a mismatch between the sentence and the previous video, participants tended to look away from the target face (post-verbally for action-verb mismatches and at the final subject region for hand gender - subject gender mismatches), suggesting outright picture-sentence incongruence can modulate the preference to inspect the face whose gender matched that of the hands seen in the previous video.</p>
      </abstract>
      <kwd-group>
        <kwd>Visual world</kwd>
        <kwd>situated sentence comprehension</kwd>
        <kwd>gender</kwd>
        <kwd>stereotypes</kwd>
        <kwd>mismatch</kwd>
        <kwd>eye tracking</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Studies on the role of world-knowledge during situated
language comprehension – i.e. when language relates to a
visual, non-linguistic context - have provided evidence that
comprehenders tend to visually anticipate contextually
plausible objects preceding their mention. In a visual
context in which a little girl, a man, a motorbike and a
carrousel were depicted, participants directed more
anticipatory looks to the motorbike after hearing The man
will ride compared to The girl will ride
        <xref ref-type="bibr" rid="ref7 ref8">(Kamide, Altmann
and Haywood, 2003)</xref>
        . Other studies have examined the
influence of prior visual context on ensuing attention and
language comprehension by showing participants a depicted
event before the sentence. What these studies have shown so
far is that visually grounded information enjoys priority
over other types of information, such as relevant linguistic
or world-knowledge. For instance, when depicted events
were pitted against world knowledge related to occupational
stereotypes, the former were preferred
        <xref ref-type="bibr" rid="ref10">(Knoeferle and
Crocker, 2007, Experiment 1)</xref>
        : During the comprehension of
non-canonical German sentences like ‘The pilotACC spies on
soon the…’ (literal translation) participants looked more
often to the (location of the) agent that had been involved in
the previously depicted spying event (e.g. a wizard) even
when a stereotypically more plausible character had also
been depicted (e.g. a detective).
      </p>
      <p>
        Manipulations of verb tense have given rise to similar
results in cases in which two objects were plausible role (i.e.
theme) fillers for the action indicated by the verb
        <xref ref-type="bibr" rid="ref10">(Knoeferle
&amp; Crocker, 2007, Experiment 3)</xref>
        . Participants saw an agent
performing an action over one of two available objects
(candelabra, crystal glasses) in a visual setting (e.g. a waiter
polishing candelabra) and then listened to a sentence such as
Der Kellner poliert… (‘The waiter polish…’). The verb was
tense ambiguous and the continuation could be about the
recent polishing event or a future polishing event (involving
the other object: the crystal glasses). Tense disambiguation
towards the simple past (i.e. …polierte soeben… ‘polished
recently’) or the future tense (i.e. …poliert demnächst…
‘polishes soon’) removed this ambiguity. Eye-movement
patterns during the verb and the adverb regions showed a
preference for the recent event target over the future event
target, regardless of tense (and before either target was
mentioned, ‘recent-event preference’). This preference has
been replicated with 5-year-old children
        <xref ref-type="bibr" rid="ref19">(Zhang, Kornbluth,
&amp; Knoeferle, 2012)</xref>
        . It also persisted in real-world settings
when the actor performed both the recent, and a “future”
event (after the sentence) in each trial
        <xref ref-type="bibr" rid="ref11">(e.g., Knoeferle,
Carminati, Abashidze, &amp; Essig, 2011)</xref>
        . A frequency bias
towards the future events (filler trials showing more
frequent post-sentential/future events) did give rise to an
earlier increase of looks to the future event targets, but the
overall recent event preference persisted
        <xref ref-type="bibr" rid="ref1 ref12 ref14">(Abashidze,
Knoeferle, &amp; Carminati, 2014)</xref>
        .
      </p>
      <p>In summary, during language comprehension, people
seem to have a strong tendency to anticipate what has
previously been visually grounded, even when this does not
match their world knowledge. The preference seems to
persist even when future tense cues should favor a future
event. However, other biases (e.g. frequency) have
modulated (but not overridden) this preference.</p>
      <sec id="sec-1-1">
        <title>Picture-sentence verification: mismatch effects</title>
        <p>
          One interesting question is to which extent the preference to
rely on visual cues (e.g., depicted events) can be modulated
by incongruence between language and the world. If we
encounter frequent mismatches and verify
world(picture)language congruence, our reliance on what is depicted might
decrease. Picture-sentence verification experiments have a
long tradition in psycholinguistics
          <xref ref-type="bibr" rid="ref5 ref6">(Gough, 1965; Just &amp;
Carpenter, 1971)</xref>
          and when combined with continuous
measures like eye tracking or ERPs (event-related brain
potentials) they can provide insights into real time language
processing
          <xref ref-type="bibr" rid="ref12 ref17 ref18">(Knoeferle, Urbach, &amp; Kutas, 2014; Vissers,
Kolk, van de Meerendonk, &amp; Chwilla, 2008; Wassernaar &amp;
Hagoort, 2007)</xref>
          .
        </p>
        <p>Wassernaar and Hagoort (2007) conducted a
picturesentence verification experiment with both healthy older
adults and aphasics. Participants inspected a line drawing
showing either a reversible event involving an agent and a
patient (e.g. a man pushing a woman) or an irreversible
event involving an agent and a theme (e.g. a woman reading
a book). Then they read a sentence in either the active (for
the semantically irreversible and reversible cases) or the
passive voice (only for the semantically irreversible cases,
e.g., ‘the tall man on this picture pushes the young woman’,
translation from Dutch). Sentences could either match or
mismatch the depicted visual information. For healthy older
participants, mismatches (vs. matches) elicited larger early
negative amplitudes time-locked to verb onset in reversible
active sentences. For the irreversible active sentences and
the reversible passive ones, the early negativity was
followed by a late positive shift, resembling a P600. The
authors argued that these ERP effects indexed processes of
thematic role assignment. Vissers and colleagues (2008)
observed a similar pattern for mismatches between the
depicted spatial relations of objects (e.g. a square followed
by a circle) and linguistically described spatial relations.
They argued in particular that the P600 was an index of a
general monitoring mechanism.</p>
        <p>
          Even more relevant to the current study,
          <xref ref-type="bibr" rid="ref12">Knoeferle et al.
(2014)</xref>
          measured ERPs as participants read English
subjectverb-object sentences (e.g. The gymnast punches the
journalist) and verified whether they matched a recently
seen picture. Sentences could either fully match the pictures
or contain different types of mismatches, i.e. verb-action,
role-relations, or both. These different mismatches elicited
distinct ERP responses. Role mismatches (vs. matches)
elicited larger anterior mean amplitude negativities
(200400 ms after subject onset) while verb-action mismatches
(vs. matches) elicited a somewhat later centro-parietal
negativity (300-500 ms after verb onset), resembling an
N400 effect. Thus, different picture-sentence relations might
actually implicate functionally distinct cognitive
mechanisms. Overall, incongruence seems to affect
language comprehension rapidly and could thus modulate
visually-based attention preferences.
        </p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>The current study</title>
      <p>The present study is motivated both by the studies on the
depicted-event preference in situated language
comprehension and by those on picture-sentence
verification. We used visual gender cues (the gender of
hands in an action video) to examine whether this sort of
visual cue (much like a recent action) would also elicit
anticipation of an upcoming depicted agent (a
gendermatching face). Participants inspected videos of pairs of
hands performing an action, and then listened to
noncanonical German object-verb-subject (OVS) sentences
while looking at the pictures of two potential agents’ faces
(one male and one female). Post-comprehension,
participants verified via a button press whether the sentence
they listened to matched the video they had just seen. We
expected faster responses for matches than mismatches.</p>
      <p>
        Expectations regarding the agent depend on the successful
association between the agents’ dimorphic gender cues from
the videotaped events (i.e. the hands) and the faces shown
during sentence comprehension. Gender categorization as
such has been characterized as straightforward
        <xref ref-type="bibr" rid="ref13 ref15 ref4">(Stangor,
Lynch, Duan, &amp; Glass, 1992; Macrae &amp; Martin, 2007;
Gaetano, van der Zwan, Blair, &amp; Brooks, 2014)</xref>
        . If this
process takes place rapidly and if gender cues influence
subsequent language comprehension, then we should
observe an early inspection preference for the target agent
face (e.g. a woman when the hands in the video belonged to
a woman) relative to the competitor agent (a man). This
anticipation would support the idea that the depicted-event
preference
        <xref ref-type="bibr" rid="ref11">(Knoeferle et al., 2011)</xref>
        generalizes to
handgender cues. The competitor however, could also receive
attention if participants base their anticipation on
stereotypical knowledge associated with the video event
(e.g., building a toy model is a stereotypically male action).
      </p>
      <p>If visual anticipation occurs, then we could furthermore
begin to examine to which extent different video-language
relations modulate it. In fact, we know little about how
different sorts of mismatches influence gender-based agency
expectations. Sentences either contained a mismatch
between the videotaped action and the sentential verb phrase
(object and verb, Experiment 1, henceforth ‘action-verb
mismatches’) or a mismatch between the gender of the
hands of the agent in the video (no face was shown in the
video) and the gender of the sentential subject (Experiment
2). Furthermore, both experiments manipulated gender
stereotypicality (i.e. the described action either
stereotypically matched or mismatched the gender of the
hands in the video).</p>
      <p>If participants rapidly relate the unfolding sentence to
their representation of previous events, both action-verb
(phrase) and hands-subject gender mismatches should
influence their expectations, resulting in attentional shifts
away from the target face and towards the competitor face.
Given that action-verb mismatches (Experiment 1) could be
detected at the sentence-initial object noun (and ensuing
verb), mismatch effects might emerge as early as the first
noun phrase or the following verb. Hands-subject gender
mismatches (Experiment 2), on the other hand, occurred at
the subject noun phrase, we should therefore observe an
effect at that region.</p>
      <p>Furthermore, if gender stereotypes have a strong influence
on visual attention during language processing, they could
potentially override or at least modulate the visual
anticipation of the target face. In that case we should
observe fewer looks to the target face in stereotypicality
mismatching cases (e.g. when female hands performed a
stereotypically male action and the sentence described a
stereotypically male action) compared to stereotype
matches.</p>
    </sec>
    <sec id="sec-3">
      <title>Experiments 1 and 2</title>
      <sec id="sec-3-1">
        <title>Participants</title>
        <p>32 different participants (16 females, 19-32 years) took part
in each experiment. Each participant received 6 Euro for
participation. All were German native speakers, had normal
or corrected to normal vision, and gave informed consent
before the experiment.</p>
      </sec>
      <sec id="sec-3-2">
        <title>Materials and Design</title>
        <p>We conducted a norming study on 104 verbalized actions
(e.g., polishing nails, repairing a radio) to assess their
gender stereotypicality. Participants (N=20) rated these on a
bipolar 7-point scale. Based on the rating results, we
selected 32 stereotypically female and 32 stereotypically
male actions as our linguistic stimuli. We paired these with
128 videos (both stereotypically female and male actions
were video-taped with a female and a male actor of which
only the hands were shown, see Fig. 1). We used two female
and two male actors. Videos were close-ups of pairs of
hands acting upon objects on a table. The target displays
bautV
(presented during sentence comprehension) consisted each
of two close-up photographs of a male and a female face.</p>
        <p>From these materials we created 32 experimental items
consisting of video pairs (one stereotypically female and
one stereotypically male action) and their corresponding
German sentence pairs with a non-canonical German
objectverb-subject (OVS) structure (see Tables 1 and 2). We used
this structure in order to monitor participants’ expectations
regarding the upcoming subject (agent). The sentences were
recorded with neutral intonation. We synchronized the
syllable length of the constituents and their onsets for the
sentences within an item.</p>
        <p>In Experiment 1 we manipulated two factors. A first was
action-verb congruence (the action described in the sentence
as expressed by object-verb combinations either matched or
mismatched the action in the video); a second factor was
stereotypicality congruence (the action described by the
sentence either matched or mismatched stereotypically with
the gender implied by the hands in the video). For instance,
a congruous condition would feature female hands in the
video and a sentence about a stereotypically female action;
an incongruous example included female hands in the video
and a sentence about a stereotypically male action.
Crossing these factors yielded 4 conditions (see Table 1),
which were counterbalanced across experimental lists in a
Latin Square manner. Word order and the use of the
postverbal adverb gleich were constant across conditions.
Experiment 2 was identical to Experiment 1 except that we</p>
      </sec>
      <sec id="sec-3-3">
        <title>Video</title>
      </sec>
      <sec id="sec-3-4">
        <title>Sentence</title>
        <p>bautV
backtV
gleichADV
gleichADV
SusannaNP2</p>
      </sec>
      <sec id="sec-3-5">
        <title>Hands subject gender match</title>
        <p>Yes
Yes
No
No
replaced the verb-action congruence manipulation with
another congruence manipulation (between the gender cued
by the hands in the video and the gender of the sentential
subject, Table 2).</p>
        <p>Fillers (total N=70) contained videos of actions which
were not stereotypically marked (e.g. filling out a
crossword), combined with the same sentence structures as
in the experimental items (N=18); videos showing two pairs
of hands engaged in an action with dative constructions
(N=18; 9 dative-first and 9 dative-middle sentences) and
pictures of objects paired with various sentence structures
(N=34). Half of the fillers contained video-sentence
mismatches of different types (e.g., action, gender of the
agents and color).</p>
      </sec>
      <sec id="sec-3-6">
        <title>Procedure</title>
        <p>An Eyelink 1000 Desktop Mounted Eye-Tracker recorded
participants’ eye movements. Viewing was binocular but we
tracked only participants’ right eye. Participants completed
10 practice trials before the experiment. Each trial started
with a video of the hands in action (3500 ms). The final
frame (displaying both hands in rest position and the object)
stayed for another 1500 ms. Next, a centrally-presented
cross appeared for 1000 ms and then a target screen with
two photographs of a female and a male face appeared (face
position was counterbalanced across trials within a list, Fig.
1). After 1500 ms the sentence was played and eye
movements to the pictures recorded. Participants responded
as quickly and accurately as possible via a (“yes” or “no”)
button press (Cedrus RB 834) whether the preceding video
matched the sentence. The position of the response buttons
was counterbalanced across participants.</p>
      </sec>
      <sec id="sec-3-7">
        <title>Analysis</title>
        <p>For each participant, we calculated the mean reaction times
- time locked to the beginning of sentence presentation - and
percentage of correct scores per condition and subjected
these to ANOVA analyses. For the eye-tracking data, we
divided each experimental sentence into four time regions,
(the object noun phrase: NP1, the verb: V; the adverb:
ADV; and the subject noun phrase: NP2). Each region
extended from its onset to the onset of the next region
except for NP2, which ended at sentence offset.</p>
        <p>
          Because looks to one of the characters implied fewer looks
to the other character in the visual scene, we computed the
mean log-gaze probability ratios for each separate sentence
region to measure the bias of inspecting the target agent (i.e.
the face which matched in gender the hands in the previous
video) over the competitor agent (the other face; ln(P(target
agent)/P(competitor)). Values above zero reflect a target
agent preference, while values below zero represent a
preference for the competitor. These scores are suitable for
parametric tests such as ANOVAs
          <xref ref-type="bibr" rid="ref3">(Arai, Van Gompel &amp;
Scheepers, 2007)</xref>
          . We calculated mean log probability ratios
per region by subjects and by items, which we subjected to
ANOVA analyses. For the time course graphs, we plotted
the gaze probability ratios in successive 20 ms time slots
from the beginning of the sentence. Missing and incorrect
responses were excluded from both the eye movement and
response time analyses.
        </p>
      </sec>
      <sec id="sec-3-8">
        <title>Time course of the eye movements</title>
        <p>The time course graphs (Fig. 2) show the time course of
eye movements to the target agent face relative to the
competitor face (ln(P(target agent)/P(competitor)) per
condition for each experiment. As no effect emerged in the
first noun phrase region (NP1), we plot the data from mean
verb onset. A first observation is that the log gaze
probability ratios were positive in all plotted time regions,
suggesting preferential inspection of the target agent (the
woman if the hands in the video had belonged to a woman)
over the competitor during the entire sentence.</p>
        <p>Upon the encounter of video-sentence mismatches a
divergence between matching (green lines) and
mismatching conditions (red lines) can be seen. While
values for matching conditions experienced an increase,
values for the mismatching conditions started to decrease,
indicating that participants started to look away from the
target face. For action-verb mismatches (Panel A) we can
see this happening at the post-verbal region, while for
hands-subject gender mismatches (Panel B), this happens at
the sentence-final subject region.</p>
      </sec>
      <sec id="sec-3-9">
        <title>Results Experiment 1</title>
        <p>Response times and accuracy: Participants were faster (ps &lt;
.001) and responded more accurately to action-verb
mismatches than matches (ps &lt; .001).</p>
        <p>Eye-movement analysis: For all regions, log ratios
remained above and significantly different from zero, which
represents the estimate of the grand mean, suggesting an
overall preference for the target face. No significant effects
of the independent variables emerged for the NP1 and V
regions. However, a significant effect of action-verb match
emerged post-verbally, at the ADV region (ps &lt; .001) and
continued into the NP2 region (ps &lt; .001): Participants
directed more looks to the target face for action-verb
matches than mismatches.</p>
      </sec>
      <sec id="sec-3-10">
        <title>Results Experiment 2</title>
        <p>Response times and accuracy: There were no significant
differences between conditions.</p>
        <p>Eye-movement analysis: Similar to Experiment 1, the log
ratios remained above zero for all regions. There was a
hands-subject match effect at NP2 (ps &lt; .01).</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Discussion</title>
      <p>In two eye-tracking verification experiments, we assessed
the generality of the depicted event preference for another
visual (gender) cue by monitoring whether the gender of
hands performing an action would affect expectations about
the agent. We used visual dimorphic gender cues and
gender-stereotype knowledge and two types of sentence
mismatches regarding the previously seen events. We varied
the video-sentence match manipulation from action-verb
(Experiment 1) to hands-subject gender congruence
(Experiment 2) and assessed in both experiments the
influence of gender stereotype mismatches. Below we report
the analyses of response latencies and accuracy in the
verification task as well as the analyses of the
eyemovement data.</p>
      <p>
        Contrary to our predictions, participants responded faster
and were more accurate for action-verb mismatches than
matches (Experiment 1). This might be due to judgment
facilitation for utterly mismatching verbal information
compared to action-verb matches, which has also been seen
in studies using a similar paradigm
        <xref ref-type="bibr" rid="ref1 ref12 ref14">(Münster, Carminati &amp;
Knoeferle, 2014, but see Knoeferle, Urbach, &amp; Kutas,
2014)</xref>
        . However, there was no such effect in Experiment 2.
It is possible that the reliable mismatch RT effect in
Experiment 1 came about because people could detect (and
thus respond to) the mismatch in principle at the first noun.
As a result, they responded faster to mismatches than
matches. In Experiment 2, the sentence-final noun phrase
revealed the hand-gender / subject-gender mismatch,
meaning that participants had to wait until sentence end
before responding to both the matches and mismatches
alike. The late emergence of the mismatching region may
have eliminated the response time differences observed in
Experiment 1.
      </p>
      <p>Regarding the eye movements, participants preferred to
look at the target face (the face whose gender matched that
of the hands in the video) relative to the competitor
throughout the sentence. This preference emerged in both
experiments and regardless of the stereotypical content of
the sentence. Importantly, video-sentence mismatches
modulated this preference, even though we did not observe
an attentional shift towards the competitor. Participants
tended to look away from the target agent when a mismatch
was encountered, suggesting that both verbal and subject
information affect expectations about the agent. This effect
emerged unexpectedly late, at the post-verbal region for
action-verb mismatches (Experiment 1) and rapidly at the
final subject region for hands-subject gender mismatches
(Experiment 2).</p>
      <p>The delayed emergence of action-verb mismatch effects
(Experiment 1) at the post-verbal region (rather than at the
first noun or verb when the mismatch could have been
detected) could indicate processes of integrating the
noncanonical object with the verb while reconciling both object
and verb with the representation of the previous event. Note
that delays in visual attention (albeit not in a mismatch
design) have been reported in studies using the same word
order (Kamide, Scheepers &amp; Altmann; 2003). Perhaps the
non-canonical structure was partly responsible for the delays
in visual attention and the action-verb congruence effect in
Experiment 1.</p>
      <p>
        Mismatch effects at the final subject region (Experiment
2), unlike action-verb mismatches, seemed more immediate,
arguably because this type of incongruence involves a single
linguistic entity. Perhaps participants adopted a referential
strategy upon encountering the subject noun. This is
compatible with the fact that they looked away from the
initially considered (target) agent in mismatching cases
(e.g., participants looked away from the male face when the
video depicted male hands but the final noun referred to a
female agent). However, a referential strategy would in
addition predict a rapid shift of attention towards the
competitor agent (e.g., the female face) closely time-locked
to its referring word
        <xref ref-type="bibr" rid="ref16">(Tanenhaus, Spivey-Knowlton,
Eberhard &amp; Sedivy, 1995)</xref>
        , which was not the case. Future
research is necessary to ascertain to which extent the
anticipation of the target face and its modulation through the
mismatches reflect referential (vs. compositional) processes.
      </p>
      <p>
        In addition, future research could examine the effects of
other mismatches and linguistic manipulations with a view
to enriching current models on situated language
comprehension
        <xref ref-type="bibr" rid="ref2 ref9">(Altmann &amp; Mirković, 2009; Knoeferle &amp;
Crocker, 2006)</xref>
        as we continue to explore the limits of visual
inspection preferences.
      </p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgements</title>
      <p>This project has received funding from the European
Union's 7th Framework Program for research, technological
development and demonstration under grant agreement
n°316748 and CITEC 277 (German Research Council).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <given-names>Abashidze D.</given-names>
            ,
            <surname>Carminati</surname>
          </string-name>
          <string-name>
            <given-names>M.N.</given-names>
            ,
            <surname>Knoeferle</surname>
          </string-name>
          <string-name>
            <surname>P.</surname>
          </string-name>
          (
          <year>2014</year>
          ).
          <article-title>How robust is the recent-event preference?</article-title>
          <source>In: Proceedings of the 36th Annual Meeting of the Cognitive Science Society</source>
          . Austin, TX: Cognitive Science Society:
          <fpage>92</fpage>
          -
          <lpage>97</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Altmann</surname>
            ,
            <given-names>G. T. M.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Mirković</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          (
          <year>2009</year>
          ).
          <article-title>Incrementality and prediction in human sentence processing</article-title>
          .
          <source>Cognitive Science</source>
          ,
          <volume>33</volume>
          ,
          <fpage>583</fpage>
          -
          <lpage>609</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>Arai</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Van</surname>
            <given-names>Gompel</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>R. P. G.</given-names>
            , &amp;
            <surname>Scheepers</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          (
          <year>2007</year>
          ).
          <article-title>Priming ditransitive structures in comprehension</article-title>
          .
          <source>Cognitive Psychology</source>
          ,
          <volume>54</volume>
          ,
          <fpage>218</fpage>
          -
          <lpage>250</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Gaetano</surname>
          </string-name>
          , J., van der Zwan, R.,
          <string-name>
            <surname>Blair</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brooks</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          (
          <year>2014</year>
          ).
          <article-title>Hands as Sex Cues: Sensitivity Measures, Male Bias Measures, and Implications for Sex Perception Mechanisms</article-title>
          .
          <source>PLoS ONE</source>
          <volume>9</volume>
          (
          <issue>3</issue>
          ):
          <fpage>e91032</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Gough P.B.</surname>
          </string-name>
          (
          <year>1965</year>
          ).
          <article-title>Grammatical transformations and speed of understanding</article-title>
          .
          <source>Journal of Verbal Learning &amp; Verbal Behavior</source>
          .
          <volume>4</volume>
          :
          <fpage>107</fpage>
          -
          <lpage>111</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Just</surname>
            ,
            <given-names>M. A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Carpenter</surname>
            ,
            <given-names>P. A.</given-names>
          </string-name>
          (
          <year>1971</year>
          ).
          <article-title>Comprehension of negation with qualification</article-title>
          .
          <source>Journal of Verbal Learning and Verbal Behavior</source>
          ,
          <volume>10</volume>
          ,
          <fpage>244</fpage>
          -
          <lpage>253</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Kamide</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Altmann</surname>
            ,
            <given-names>G. T. M.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Haywood</surname>
            ,
            <given-names>S. L.</given-names>
          </string-name>
          (
          <year>2003</year>
          ).
          <article-title>The time-course of prediction in incremental sentence processing: Evidence from anticipatory eye movements</article-title>
          .
          <source>Journal of Memory and Language</source>
          ,
          <volume>49</volume>
          ,
          <fpage>133</fpage>
          -
          <lpage>159</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Kamide</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Scheepers</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Altmann</surname>
            ,
            <given-names>G.T.M.</given-names>
          </string-name>
          (
          <year>2003</year>
          ).
          <article-title>Integration of syntactic and semantic information in predictive processing: Cross-linguistic evidence from German and English</article-title>
          .
          <source>Journal of Psycholinguistic Research</source>
          ,
          <volume>32</volume>
          ,
          <fpage>37</fpage>
          -
          <lpage>55</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Knoeferle</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Crocker</surname>
            ,
            <given-names>M. W.</given-names>
          </string-name>
          (
          <year>2006</year>
          ).
          <article-title>The coordinated interplay of scene, utterance, and world knowledge: Evidence from eye-tracking</article-title>
          .
          <source>Cognitive Science</source>
          ,
          <volume>30</volume>
          ,
          <fpage>481</fpage>
          -
          <lpage>529</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>Knoeferle</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          &amp;
          <string-name>
            <surname>Crocker</surname>
            ,
            <given-names>M. W.</given-names>
          </string-name>
          (
          <year>2007</year>
          ).
          <article-title>The influence of recent scene events on spoken comprehension: Evidence from eye movements</article-title>
          .
          <source>JML</source>
          ,
          <volume>57</volume>
          (
          <issue>4</issue>
          ):
          <fpage>519</fpage>
          -
          <lpage>543</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>Knoeferle</surname>
            <given-names>P</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Carminati</surname>
            <given-names>MN</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abashidze</surname>
            <given-names>D</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Essig</surname>
            <given-names>K</given-names>
          </string-name>
          (
          <year>2011</year>
          ).
          <article-title>Preferential inspection of recent real-world events over future events: evidence from eye tracking during spoken sentence comprehension</article-title>
          .
          <source>Frontiers in Psychology</source>
          <volume>2</volume>
          :
          <fpage>306</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <surname>Knoeferle</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Urbach</surname>
            ,
            <given-names>T.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kutas</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          (
          <year>2014</year>
          ).
          <article-title>Different mechanisms for role relations versus verb-action congruence effects: Evidence from ERPs in picturesentence verification</article-title>
          ,
          <source>Acta Psychologica</source>
          , vol.
          <volume>152</volume>
          , pp.
          <fpage>133</fpage>
          -
          <lpage>148</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <surname>Macrae</surname>
            ,
            <given-names>C. N.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Martin</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          (
          <year>2007</year>
          ).
          <article-title>A boy primed sue: Feature-based processing and person construal</article-title>
          .
          <source>European Journal of Social Psychology</source>
          ,
          <volume>37</volume>
          ,
          <fpage>793</fpage>
          -
          <lpage>805</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <surname>Münster</surname>
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Carminati</surname>
            <given-names>M.N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Knoeferle</surname>
            <given-names>P.</given-names>
          </string-name>
          (
          <year>2014</year>
          ).
          <article-title>How Do Static and Dynamic Emotional Faces Prime Incremental Semantic Interpretation?: Comparing Older and Younger Adults</article-title>
          .
          <source>In: Proceedings of the 36th Annual Meeting of the Cognitive Science Society</source>
          . Austin, TX: Cognitive Science Society;
          <year>2014</year>
          :
          <fpage>2675</fpage>
          -
          <lpage>2680</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <surname>Stangor</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lynch</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Duan</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Glass</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          (
          <year>1992</year>
          ).
          <article-title>Categorization of individuals on the basis of multiple social features</article-title>
          .
          <source>J Pers Soc Psychol</source>
          <volume>62</volume>
          :
          <fpage>207</fpage>
          -
          <lpage>218</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <surname>Tanenhaus</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Spivey-Knowlton</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Eberhard</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Sedivy</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          (
          <year>1995</year>
          ).
          <article-title>Integration of visual and linguistic information during spoken language comprehension</article-title>
          .
          <source>Science</source>
          ,
          <volume>268</volume>
          ,
          <fpage>1632</fpage>
          -
          <lpage>1634</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <surname>Vissers</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kolk</surname>
            , H., van de Meerendonk,
            <given-names>N.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Chwilla</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          (
          <year>2008</year>
          ).
          <article-title>Monitoring in language perception: Evidence from ERPs in a picture-sentence matching task</article-title>
          .
          <source>Neuropsychologia</source>
          ,
          <volume>967</volume>
          -
          <fpage>982</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <surname>Wassenaar</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Hagoort</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          (
          <year>2007</year>
          ).
          <article-title>Thematic role assignment in patients with Broca' aphasia: Sentence picture matching electrified</article-title>
          .
          <source>Neuropsychologia</source>
          ,
          <volume>716</volume>
          -
          <fpage>740</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kornbluth</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Knoeferle</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          (
          <year>2012</year>
          ).
          <article-title>The role of recent versus future events in childand adult language comprehension: Evidence from eye tracking. Poster presented in the 18th Annual Conference on Architectures and Mechanisms for Language Processing</article-title>
          (AMLaP),
          <source>Riva del Garda</source>
          , Italy
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>