<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Analysis of Gaze Tra jectories in Natural Reading with Hidden Markov Models</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Lomonosov Moscow State University</institution>
        </aff>
      </contrib-group>
      <fpage>419</fpage>
      <lpage>428</lpage>
      <abstract>
        <p>The process of natural reading, registered with a modern eye-tracking system generates a signal of complicated structure that can be considered as a time series consisting of gaze point coordinates. Signal properties are supposed to depend on various properties of presented text as well as on current cognitive condition of a reader such as attention focus, level of fatigue, level of text understanding and other parameters. The task of cognitive state recognition can be approached with the modeling of gaze trajectories using probabilistic models, which parameters may contain information relevant to read text properties and reader's cognitive state. In this work a new approach of gaze trajectories modeling based on Hidden Markov Models is proposed. HMM's transition probability matrix corresponds to probabilities of saccades between words and emission probability functions correspond to words coordinates and overall measurement noise. Two variants of HMM are proposed: text-related HMM models multiple gaze trajectories collected on the same text from di erent readers, subject-related parametric HMM models gaze trajectories produced by a single reader on a set of consecutive pages from the same text. A series of experiments on simulated data were performed to estimate a required sample size and a required level of measurement accuracy for a forthcoming data collection procedure.</p>
      </abstract>
      <kwd-group>
        <kwd>Eye-tracking</kwd>
        <kwd>Natural reading</kwd>
        <kwd>Hidden Markov Models</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Natural reading is a complex task that includes eye movement processes,
lexicosemantic processing dependent on reader attention and visual features of the
text being read.</p>
      <p>The process of reading consists of long relatively rare movements ("saccades")
between areas of high attention where eyes are xated on a word for some time
depending on the skill of reader.</p>
      <p>
        Eye movements during reading are under the direct control of linguistic
processing [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. There are three properties of a word that in uence its ease of
processing: word's frequency, length and predictability in context, the so-called Big
Three [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. These words properties a ect time of word's processing, length of
saccade and probability of word's skipping. Thereby longer words require more
processing time from reader because they consist of larger number of letters
then shorter, so they include more information to process. Although the word's
frequency is correlated with it's length in general case, it has been shown that
more frequent words are processed faster then infrequent words with similar
lengths [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Moreover, word's length a ects not only the duration of current
xation but the length of next saccade and duration of next xation [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ]. More
predictable words are more likely to be skipped and require shorter xations.
      </p>
      <p>
        Medical conditions also in uence eye-movements during reading. For
example, schizophrenia patients read slower, make shorter saccades, have longer
forward xation durations and make a greater number of regressive saccades [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
Several studies show that children with dyslexia read slower, make a larger
number of long progressive saccades, make twice more xations on long words and
less often skip short words then children without this condition [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ].
      </p>
      <p>
        Eye-tracking signals are successfully used in a large number of di erent tasks.
Subject's attention during reading can be determined with respect to type of
reading: reading, skimming and scanning based on features extracted from
eyetracking data: amplitude, angularity, velocity of saccades and duration of
xations [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Eye-movements data may be used in tasks of language pro ciency
determining [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. Students' eye movements taken while IELTS reading test
completion have been analyzed and signi cant di erence between samples taken from
successful and unsuccessful attempts to pass the exam has been revealed [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. In
a part-of-speech tagging task data extracted from eye-tracker allows to improve
the performance of an approach based on Second-order Hidden Markov Models.
The extracted data was transformed into 22 features encoding information about
xation durations, probabilities, number of xations, re xations and regressions
related to current word and its neighbours [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. The same features are found to
be useful in a domain of Named Entity Recognition in approach based on a usage
of a bidirectional LSTM. Using embeddings based on these features it became
possible to improve performance of the previous state-of-the art model [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>
        An approach based on applying Slip-Kalman ltering is used to track the
progression of reading. This approach works particularly well for determining
the event of changing a read line but also shows good results with respect to
noise reduction [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
      </p>
      <p>At the works discussed above di erent variations of Hidden Markov Models
are used in order to determine the part of speech, the correct coordinate at
which the eye is directed etc. In this work we propose to model eye-tracking
trajectories using Hidden Markov Models: associate a set of hidden states with
individual read words and t the matrix of transition probabilities between them.
We assume that this matrix can be used as a set of features for a models solving
various cognitive state recognition tasks.
2
2.1</p>
    </sec>
    <sec id="sec-2">
      <title>Proposed Approach</title>
      <sec id="sec-2-1">
        <title>Hidden Markov Models</title>
        <p>A rst-order Hidden Markov Model is a probabilistic model which is based on
Markov Chain model. An HMM is de ned by the following components:
{ X = fx1; : : : ; xN g; xn 2 Rd { a sequence of observed values</p>
        <p>K
{ T = ft1; : : : ; tN g; tn 2 f0; 1gK ; P tnj = 1 { a sequence of hidden states that
j=1
ti =
correspond to observed values.</p>
        <p>(1; if a model in a state i</p>
        <p>0; otherwise
{ A transition probability matrix AK K , where aij = p(tn;j jtn 1;i) is a
probak
bility of transition from the hidden state tn 1;i to the state tnj , P aij = 1 8i
j=1
{ p (xnjtn) { observation likelihoods or emission probabilities, expressing
probabilities that observed value xn would be generated from a hidden state tn. It
is assumed that conditional distribution p (xnjtn) is known up to parameters
k; k 2 f1; : : : ; Kg, so if tni = 1 then xn is from p(xnj i)
{ = f 1; : : : ; K g { an initial probability distribution. i is the probability</p>
        <p>N
of HMM being started in state i. P i = 1</p>
        <p>j=1
A rst-order HMM instantiates 2 assumptions. The rst one is Markov
assumption: the value of the hidden state tn depends only on the state at the previous
moment tn 1.</p>
        <p>p(tnjt1; : : : ; tn 1) = p(tnjtn 1)
Second, the value of the observed value xn depends only on the current hidden
state. It is known as Output Independence assumption.</p>
        <p>p(xnjX; T ) = p(xnjtn)
(1)
(2)</p>
        <p>Having a sequences of observed values (in our case, coordinates taken from an
eye-tracker) we need to determine from which hidden states (read words) these
coordinates were led from. It is needed to solve HMM learning problem: learn the
transition probability matrix A and observation likelihoods p (xnjtn) from given
observation sequences X and set of possible hidden states where each hidden
state would represent read word. Let's denote the set of HMM parameters as =
f ; A; g. For an HMM model parameters estimation Expectation-Maximization
or EM algorithm can be applied.</p>
        <p>= arg max p(Xj ) = arg max X p(X; T j )</p>
        <p>T
p(Xj ) =</p>
        <p>X p(X; T j ) ! max , log
T</p>
        <p>X p(X; T j )
T
1. Initialization step. At the beginning it is needed to set
eters.</p>
        <p>{</p>
        <p>K
P
j=1
and A are usually set randomly but corresponding to restrictions
k
i = 1 and P aij = 1 8i
j=1
!</p>
        <p>! max
= f ; A; g
param{ initialization depends on p(xj ) distributions
2. Expectation step. old are xed.</p>
        <p>ET jX; old log p(X; T j ) =</p>
        <p>X log p(X; T j )p(T jX; old)
T
3. Maximization step. p(T jX; old) are xed.</p>
        <p>new = arg max ET jX; old log p(X; T j )
4. Expectation and Maximization steps are repeated until convergence</p>
        <p>Baum-Welch algorithm is the special case of the standard HMM-training
algorithm which allows to perform E and M steps more e ciently due to
optimization related to HMM assumptions 1 and 2.</p>
        <p>Our primary goal is to propose an approach for extracting subject-dependent
features from eye-tracking data that would correspond to cognitive states and
a group of text-dependent features. Examples of cognitive states that can be
extracted are a level of fatigue, stress, focus, a level of text understanding,
emotional state. One of the features of the text can be "di cult" words that take
a longer reading time, require higher number of xations and longer xations
that can be explained by the requirements for the level of pro ciency in the
language or subject area. Or we can try to discern the in uence of words which are
unpredictable in context on eyes movements patterns. Eye-tracking data can be
represented as a sequence of coordinates taken from eye-tracker with resolution
of several hundred coordinates per second. Each coordinate from a sequence can
be assigned to an area of high attention such as a read word. It is proposed to
learn an HMM using coordinates as observed values and determine the set of
hidden states based on knowledge about the number of words. From practical
point of view, one of two following situations is considered for further analysis:
either a same text is read by a certain number of di erent subjects, or a
single subject reads a set of texts of su cient size. Thus, two di erent models are
proposed for these scenarios, text-dependent and subject-dependent.
2.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>Task 1: Text Analysis</title>
        <p>In a "same text, di erent readers" scenario, text-dependent features remain the
same for every session, but reader-dependent features should be di erent for
di erent readers.</p>
        <p>The main objective in the analysis of this scenario is to determine a vector
of text-dependent parameters given a set of observation sequences related to
one text fragment read by a certain number of subjects. It is assumed that
a set of observations related to a large enough set of subjects can be used to
estimate a subject-independent HMM parameters which can be analysed for
the purpose of extracting text-related features while subject-dependent features
would be suppressed. Thus the text-dependent HMM consists of the following
components:
{ Xj = fx1j ; : : : ; xnj g { is a set of observed values taken from subject j.
{ T = ft1; : : : ; tK g { is a set of hidden states determine by a number of words
in the text.
{ p(xnjtn) { emission probabilities { set of two-dimensional Gaussian
distributions
N = fN1( 1; 1); : : : ; Nn( n; n)g with means de ned by geometric centers
of words C = fc1; : : : ; cK g, i = ci and standard deviations represents
overall noise of measurements of the eye-tracking device and is estimated
from the data.
{ = f 1; : : : ; K g { initial probability distribution. It also can be estimated
from samples.
{ AK K { a transition probability matrix that can be tted from set of
eyetracking measurements taken from di erent subjects.</p>
        <p>Parameters of the model trained from samples taken from di erent subjects
reading the same text may potentially represent such text characteristics as
sentence structure, word frequency, general text complexity, etc.
2.3</p>
      </sec>
      <sec id="sec-2-3">
        <title>Task 2: Subject Analysis</title>
        <p>Let's consider a scenario when a single subjects reads a certain number of text
fragments. It may be assumed that during one session cognitive states of a
subject should not change signi cantly over time and therefore model parameters
should be similar for every single page. The objective of the analysis in this
scenario is to estimate these subject-dependent parameters. Since the data is
presented as a set of eye-tracker trajectories collected from di erent text
fragment, each single trajectory is assumed to be sampled from a corresponding
text-dependent HMM. It is assumed that parameters of each HMM can be
presented as a function of higher-level parameters that refer to current cognitive
state of a reader. For example, a level of fatigue can be modeled as the average
duration of a xation on a word. In order to train a subject-dependent HMM on
these data, a parametric family of HMM is proposed.</p>
        <p>Suppose that is a vector of parameters that represents a cognitive state of a
subject. Such parameters as average number of xations, average xation
duration, average number of regressions could be represented as functions dependent
on the vector and words frequencies. In a simplest case, a parametric family
of HMM can be proposed in a following form:
{ M = fm1; : : : ; mP g is a set of HMMs, mi is a HMM corresponding to text
i.
{ For each HMM emission probabilities are de ned by two-dimensional
Gaussian distributions with parameters considering geometry of text and
measurements noise.
{ For each HMM a set of hidden states is determined by number of words
W = fw1; : : : ; wKp g presented in a text fragment.
{ Reading process is modeled using probabilities ( ; W ), ( ; W ), ( ; W ),
( ; W ), where ( ; W ) is a probability of continuing a current xation,
( ; W ) is a probability of saccade to a next word, ( ; W ) is a probability
of saccade to a previous word, ( ; W ) is a probability of long forward or
backward saccade.</p>
        <p>Thus a transition probability matrices have the following form:
0</p>
        <p>( ; w1)
B ( ; w2)</p>
        <p>BB ( ; w3)
AK K = BBB ...</p>
        <p>B@ ( ; wK 1)
( ; wK )
( ; w1) ( ; w1) : : : : : : ( ; w1)
( ; w2) ( ; w2) : : : : : : ( ; w2) C
( ; w3) ( ; w3) : : : : : : ( ; w3) CC
... ... . . . ... ... CCC
: : : : : : : : : ( ; wK 1) ( ; wK 1) AC
: : : : : : : : : ( ; wK ) ( ; wK )
1
Thus, a parametric form of subject-dependent HMM can be de ned by setting a
vector of parameters and choosing a set of functions ; ; ; . The model can
be more detailed if other text-related parameters would be taken into account.
3</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Eye-tracking Corpora</title>
      <p>
        For future studies and experiments two existing eye-tracking datasets were
chosen: Zurich Cognitive Language Processing Corpus [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] and Ghent Eye-Tracking
Corpus [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ].
      </p>
      <p>ZuCo is a publicly available dataset containing eye-tracking and EEG data
recorded from 12 native English speakers reading natural English text. Subjects
were presented with three tasks: normal reading task in which participants had
to give an assessment of movie described in a read text fragment, normal reading
task with multiple choice questions about read content and task-speci c reading
task in which subjects had to focus on a certain semantic relation type. The
corpus includes high density eye-tracking data recorded with a calibrated infrared
eye-tracker. Fixations, saccades and blinks are identi ed by the tracker software.
The dataset also includes such statistics of gaze trajectories as time after rst
xation on sentence for every single xation, number of xations for each word
and sentence and mean pupil size, gaze duration (GD) during rst word reading,
sum of reading time of word (total reading time or TRT), rst xation duration
(FFD), the duration of rst and single xation on a word (single xation
duration or SFD) and go-past time (GPD) which is the sum of all xations preceding
saccade to the right.</p>
      <p>GeCo is a corpus of eye-tracking data taken from 14 monolingual and 19
bilingual participants reading a novel. Bilinguals were classi ed as English speakers
with pro ciency level from lower-intermediate to advanced. Bilingual
participants read half of the novel in their rst language and the other half in their
second language. The size of the text read by each subject was about 5,000
sentences. As in ZuCo dataset in addition to raw data extracted by the tracker
there are presented word-level reading measurements: GD, FFD, SFD, TRT and
GPT.
4</p>
    </sec>
    <sec id="sec-4">
      <title>Experiments</title>
      <p>A set of experiments was executed to prove a concept that eye-tracking
trajectories can be generated using HMM transition probability matrix and then HMM
can be tted the model parameters can be tted close to the parameters of the
original model. For a given set of sentences we can obtain coordinates of exact
positions of words on a page. The distance was measured in relation to the size
of a printed letter. A Hidden Markov Model was initialized in the following way
according to our vision of what they should look like. The number of hidden
states was set according to the number of displayed words. A transition
probability matrix was generated as diagonally dominant since the number of saccades
and therefore hidden states transitions has to be much smaller then number of
eye movements inside areas of gaze xations on a single word when the hidden
state does not change. Initial probability distribution was chosen to be geometric
distribution in order to simulate a task in which the subject must read the text
from the beginning. A set of gaussian 2-d distributions with means located in
word centers was chosen as a set of emission probabilities.</p>
      <p>A simulated training dataset of sequences of observed values was generated
using the de ned HMM. For a texts with several dozen words each sequence
consists of 300 observed values. We perform an experiment to nd out how many
observation is needed for a precise tting of HMM parameters. The process of
model training was run for datasets with di erent number of observations. For
each size of training sample a new model was trained. Each model was trained
using its own dataset. Then mean squared error between original transition
probability matrix and trained transition probability matrix and mean squared error
between main diagonals of original and trained matrices for each trained model
were measured.</p>
      <p>MSE =</p>
      <p>K
1</p>
      <p>K K</p>
      <p>X X(Aij</p>
      <p>K i=1 j=1
MSE diagonal =
1 K</p>
      <p>X(Aii
K i=1</p>
      <p>Aij )2
Aii)2
(3)
(4)</p>
      <p>N 1
s =
MSE mean values and their 95% con dence intervals were also calculated.
Condence interval were de ned as x Q(1 2 ) s where x is the sample mean,
s N
( P xi x)2
i=1
is the sample standard deviation,</p>
      <p>is the con dence level, N
0.012
0.011
0.010
0.009
E
SM0.008
0.007
0.006
0.005
0.004
0.200
0.175
lan0.150
o
g
a
id0.125
E
S
M0.100
0.075
0.050
0
5
10
is the sample size, Q is quantile nction: Q(p) = inffx 2 R : p F (x)g, F (x) is
the cumulative distribution function for the Student's t-distribution with N 1
degrees of freedom. A con dence interval gives a more descriptive estimate of a
parameter than a point estimation. Results of our experiments are presented on
Fig. 1 and Fig. 2.</p>
      <p>
        The experiments were executed using python package named hmmlearn [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ].
As is shown in the gures our metrics reach values close to optimal at
observations number between 10 and 15. With a further increase in observations numbers
MSE over full transition probability matrices and MSE over diagonals do not
decrease signi cantly. So for a suboptimal quality it could be enough to collect
data from dozen of subjects.
5
      </p>
    </sec>
    <sec id="sec-5">
      <title>Conclusions and Future Work</title>
      <p>In this work various approaches to the analysis of eye movement in di erent
text-dependent and subject-dependent scenarios have been considered. A new
approach of gaze trajectories modeling based on Hidden Markov Models is
proposed for both scenarios. An experiment conducted as part of a preliminary
study helped us determine the approximate size of the sample we would need for
a further work. It is planned to apply a proposed approach on real data taken
from ZuCo and GeCo datasets in one of the following tasks.
1. Binary classi cation task: was the answer given by subject in a sentiment
task from ZuCo dataset was correct or not.
2. If information about the answer time is available it may be possible to
evaluate this parameter using a model, because it is supposed that time required
for answer is dependent on how thoroughly the text was read.
3. The fatigue classi cation task. It is assumed that at the end of the long
reading session a fatigue level is higher then at the beginning of the session.</p>
      <p>Data from GeCo corpus would be useful.</p>
      <p>It is also planned to collect a new dataset containing samples of eye movement
recordings taken while reading text fragments in Russian. Several dozen subjects
will read several texts at di erent levels of fatigue. Fatigue level will be measured
in two ways: by interviewing subjects and using a binary classi er trained to
recognize the beginning and end of the session. Using the second method, we
will assume that fatigue level rises at the end of the session.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Dambacher</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Slattery</surname>
            ,
            <given-names>T. J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kliegl</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rayner</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <article-title>Evidence for direct control of eye movements during reading</article-title>
          .
          <source>Journal of Experimental Psychology: Human Perception and Performance</source>
          ,
          <volume>39</volume>
          (
          <issue>5</issue>
          ):
          <volume>1468</volume>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>Clifton</given-names>
            <surname>Jr</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            ,
            <surname>Ferreira</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            ,
            <surname>Henderson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            ,
            <surname>Inho</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. W.</given-names>
            ,
            <surname>Liversedge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. P.</given-names>
            ,
            <surname>Reichle</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. D.</given-names>
            ,
            <surname>Schotter</surname>
          </string-name>
          ,
          <string-name>
            <surname>E. R.</surname>
          </string-name>
          <article-title>Eye movements in reading and information processing: Keith Rayner's 40 year legacy</article-title>
          .
          <source>Journal of Memory and Language</source>
          ,
          <volume>86</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>19</lpage>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Rayner</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          , Du y, S. A.
          <article-title>Lexical complexity and xation times in reading: E ects of word frequency, verb complexity, and lexical ambiguity</article-title>
          .
          <source>Memory &amp; cognition 14</source>
          .3:
          <fpage>191</fpage>
          -
          <lpage>201</lpage>
          (
          <year>1986</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>White</surname>
            ,
            <given-names>S. J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rayner</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liversedge</surname>
            ,
            <given-names>S. P.</given-names>
          </string-name>
          <article-title>The in uence of parafoveal word length and contextual constraint on xation durations and word skipping in reading</article-title>
          .
          <source>Psychonomic bulletin &amp; review 12</source>
          .3:
          <fpage>466</fpage>
          -
          <lpage>471</lpage>
          (
          <year>2005</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Juhasz</surname>
            ,
            <given-names>B. J.</given-names>
          </string-name>
          , et al.
          <article-title>Eye movements and the use of parafoveal word length information in reading</article-title>
          .
          <source>Journal of Experimental Psychology: Human Perception and Performance</source>
          <volume>34</volume>
          .6:
          <issue>1560</issue>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Whitford</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          , et. al.
          <article-title>Reading impairments in schizophrenia relate to individual differences in phonological processing and oculomotor control: Evidence from a gazecontingent moving window paradigm</article-title>
          .
          <source>Journal of Experimental Psychology: General</source>
          <volume>142</volume>
          .1:
          <issue>57</issue>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Olson</surname>
            ,
            <given-names>R. K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kliegl</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Davidson</surname>
            ,
            <given-names>B. J.</given-names>
          </string-name>
          <article-title>Dyslexic and normal readers' eye movements</article-title>
          .
          <source>Journal of Experimental Psychology: Human Perception and Performance 9</source>
          .5:
          <issue>816</issue>
          (
          <year>1983</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>De Luca</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , et. al.
          <article-title>Eye movement patterns in linguistic and non-linguistic tasks in developmental surface dyslexia</article-title>
          .
          <source>Neuropsychologia</source>
          <volume>37</volume>
          .12:
          <fpage>1407</fpage>
          -
          <lpage>1420</lpage>
          (
          <year>1999</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9. Moza ari, S. S. et al.
          <article-title>Reading Type Classi cation based on Generative Models and Bidirectional Long Short-Term Memory</article-title>
          .
          <source>Joint Proceedings of the ACM IUI 2018 Workshops. CEUR Workshop Proceedings</source>
          <year>2068</year>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Kunze</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          , et. al.
          <article-title>I know what you are reading: recognition of document types using mobile eye tracking</article-title>
          .
          <source>Proceedings of the 2013 International Symposium on Wearable Computers</source>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Kunze</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          et. al.
          <article-title>Towards inferring language expertise using eye tracking</article-title>
          .
          <source>CHI'13 Extended Abstracts on Human Factors in Computing Systems</source>
          ,
          <volume>217</volume>
          -
          <fpage>222</fpage>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Bax</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <article-title>Readers' cognitive processes during IELTS reading tests: Evidence from eye tracking</article-title>
          .
          <source>British Council, ELT Research Papers 13-06</source>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Barrett</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          , et. al.
          <article-title>Weakly supervised part-of-speech tagging using eye-tracking data</article-title>
          .
          <source>Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics</source>
          , Volume
          <volume>2</volume>
          :
          <string-name>
            <given-names>Short</given-names>
            <surname>Papers</surname>
          </string-name>
          (
          <year>2016</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Hollenstein</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhang</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <article-title>Entity Recognition at First Sight: Improving NER with Eye Movement Information</article-title>
          . arXiv preprint arXiv:
          <year>1902</year>
          .
          <volume>10068</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Bottos</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Balasingam</surname>
            ,
            <given-names>B. A Novel</given-names>
          </string-name>
          <string-name>
            <surname>Slip-Kalman</surname>
          </string-name>
          <article-title>Filter to Track the Progression of Reading Through Eye-Gaze Measurements</article-title>
          . arXiv preprint arXiv:
          <year>1907</year>
          .
          <volume>07232</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Hollenstein</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          , et al.
          <article-title>ZuCo, a simultaneous EEG and eye-tracking resource for natural sentence reading</article-title>
          .
          <source>Scienti c data 5</source>
          .1:
          <fpage>1</fpage>
          -
          <lpage>13</lpage>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Cop</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          , et al.
          <article-title>Presenting GECO: An eyetracking corpus of monolingual and bilingual sentence reading</article-title>
          .
          <source>Behavior research methods 49</source>
          .2:
          <fpage>602</fpage>
          -
          <lpage>615</lpage>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>18. hmmlearn library. https://hmmlearn.readthedocs.io</mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Ulutas</surname>
            ,
            <given-names>B. H.</given-names>
          </string-name>
          , Ozkan,
          <string-name>
            <surname>N. F.</surname>
          </string-name>
          , Michalski, R.
          <article-title>Application of hidden Markov models to eye tracking data analysis of visual quality inspection operations</article-title>
          .
          <source>Central European Journal of Operations Research</source>
          ,
          <volume>28</volume>
          :
          <fpage>761</fpage>
          {
          <fpage>777</fpage>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>