<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Isabel Segura-Bedmar</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Luis N u´nez-G o´mez</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Paloma Mart´ınez</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Maribel Quiroz</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Computer Science Department, University Carlos III of Madrid Avd. Universidad</institution>
          ,
          <addr-line>30,Legane ́s, Madrid</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Drug Package Leaflets provide information for patients on how to safely use medicines. European Commission and recent studies stress that further efforts must be made to improve the readability and understandability of package leaflets in order to ensure the proper use of medicines and to increase patient safety. To the best of our knowledge, this is the first work that directly deals with the automatic simplification of drug package leaflets. Our approach to lexical simplification combines the use of domain terminological resources to give a set of synonym candidates for a given target term, and the use of their frequencies in a large collection of documents in order to select the simplest synonym.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>
        Since 2001, according to a directive of the
European Parliament
        <xref ref-type="bibr" rid="ref16">(Directive 2001/83/EC)</xref>
        <xref ref-type="bibr" rid="ref16">(EU,
2001)</xref>
        , every drug product must be accompanied
by a package leaflet before being placed on the
market. This document provides informative
details about a medicine, including its appearance,
actions, side effects and drug interactions,
contraindications, special warnings, etc. This
directive also required that Drug Package Leaflets
(DPL) must be written in order to provide clear
and comprehensible information for patients since
their misunderstanding could be a potential source
of drug related problems, such as medication
errors and adverse drug reactions. In 2009, the
European Commission published a guideline
        <xref ref-type="bibr" rid="ref14">(EC,
2009)</xref>
        with recommendations and advices in order
to issue package leaflets with accessible and
understandable information for patients. However,
recent studies
        <xref ref-type="bibr" rid="ref32 ref33">(Pires et al., 2015; Pin˜ero-Lo´pez
et al., 2016)</xref>
        show that the readability and
understandability of these documents have not been
improved during the last seven years. Therefore,
further efforts must be made to improve the
understandability of package leaflets in order to ensure
the proper use of medicines and to increase patient
safety.
      </p>
      <p>
        One of the main reasons why the
understandability has not been improved is that these
documents still contain a considerable number of
technical terms describing adverse drug reactions,
diseases and other medical concepts. Posology
(dosage quantity and prescription),
contraindications and adverse drug reactions seem to be the
sections most difficult to understand
        <xref ref-type="bibr" rid="ref25">(March et al.,
2010)</xref>
        . To help solving this problem, we
propose an automatic system to simplify drug
package leaflets.
      </p>
      <p>Text simplification is a Natural Language
Processing (NLP) task that aims to rewrite text into
an equivalent with less complexity for readers.
There are two main approaches to this task: lexical
and syntactic simplification. Lexical
simplification basically consists of replacing complex
concepts with simpler synonyms, while syntactic
simplification aims to reduce the grammatical
complexity of a text while preserving its meaning.</p>
      <p>
        Text simplification techniques have been
applied to simplify texts from different domains such
as crisis management
        <xref ref-type="bibr" rid="ref41">(Temnikova, 2012)</xref>
        , health
information
        <xref ref-type="bibr" rid="ref19 ref20 ref21">(Jonnalagadda et al., 2009; Kandula
et al., 2010; Jonnalagadda and Gonzalez, 2011)</xref>
        ,
aphasic readers
        <xref ref-type="bibr" rid="ref11">(Devlin, 1999)</xref>
        , language
learners
        <xref ref-type="bibr" rid="ref26 ref31">(Petersen and Ostendorf, 2007)</xref>
        .
Comprehensive surveys of the text simplification field can be
found in
        <xref ref-type="bibr" rid="ref37 ref39">(Shardlow, 2014; Siddharthan, 2014)</xref>
        .
      </p>
      <p>To the best of our knowledge, this is the first
work that directly deals with the automatic
simplification of drug package leaflets. In particular,
we focus on the lexical simplification of adverse
drug reactions that are described in these
documents. Moreover, our work is one of the few
studies that address the simplification of texts written
in Spanish. Our approach for lexical simplification
combines the use of terminological resources that
provide a set of synonym candidates for a given
target term, and the use of their frequencies in a
large collection of documents in order to select the
most common synonym.</p>
      <p>The paper is organized as follows. Section 2
presents related work. Section 3 describes our
approach. Experiments, results, and discussion are
given in Section 4. Finally, the paper is concluded
and future work is proposed in Section 5.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        First works in text simplification started 20 years
ago
        <xref ref-type="bibr" rid="ref6">(Chandrasekar et al., 1996)</xref>
        . It is based on
transforming a text in an equivalent text that is
easier to read and probably easier to understand by a
target audience.
      </p>
      <p>
        There is a need to adapt contents for some
groups of people because information is not
equally accessible to everyone. It is unlikely that
professional editors will adapt text for all
literacy levels, and NLP techniques could help
simplify texts by automating some tasks. In this way,
it is possible to help content editors to generate
adapted contents. On the other hand, text
simplification is essential in several types of texts: News,
Government and administrative information, laws
and rights, etc. As it was mentioned before, there
are two subtasks of text simplification
        <xref ref-type="bibr" rid="ref34">(Saggion
et al., 2011)</xref>
        : (1) syntactic simplification that
divides complex sentences in simplest sentences, (2)
lexical simplification whose objective is to
substitute complex vocabulary by common vocabulary
(looking for synonyms that are simpler than the
original word considering the context in the
sentence). Moreover, a clarification step could be
included to provide definitions and explanations
for acronyms, abbreviations and unusual words.
These tasks are not completely automatic, they
have to be manually reviewed in some cases.
      </p>
      <p>
        Firstly, we have to distinguish between
readability and understandability because these
concepts capture different aspects of the complexity
of the text. Readability is about the structure of
sentences (it concerns syntax and consequently
requires syntactic simplification approaches). On
the other side, understandability is about the
difficulty to interpret a word
        <xref ref-type="bibr" rid="ref3">(Barbieri et al., 2005)</xref>
        and lexical simplification approaches are required.
      </p>
      <p>
        Concerning syntactic simplification it consists
on transforming complex and long sentences into
simplest and independent sentences eliminating
coordination (of clauses, verbs, etc.), dropping
subordination utterances (relative clauses,
gerundive and participle utterances), resolving anaphora
and transforming passive into active voice. First
a parser is used to obtain a dependency tree that
represents the syntactic structure of the sentence
(noun, prepositional and verbal phrases and how
they are related to)
        <xref ref-type="bibr" rid="ref12">(Dorr et al., 2003)</xref>
        . Then,
rule-based approaches are used in syntactic
simplification. Rules can be automatically learned
from annotated corpora of text (syntactic trees of
sentences where original sentences are related to
their simplified sentences)
        <xref ref-type="bibr" rid="ref42">(Zhu et al., 2010)</xref>
        , or
handcrafted rules
        <xref ref-type="bibr" rid="ref38 ref6">(Chandrasekar et al., 1996;
Siddharthan, 2002)</xref>
        . The rules include split, drop,
copying and reordering operations over syntactic
trees.
      </p>
      <p>
        Related to lexical simplification, this task
consists on replacing words (taking into account the
context) and complex utterances by easier words
or phrases. A heuristic used is that complex words
have a low frequency. Moreover, lexical resources,
as Wordnet
        <xref ref-type="bibr" rid="ref28">(Miller, 1995)</xref>
        , are used to extract
synonyms as candidates to replace a complex or
difficult word. Combining a lexical resource and a
probabilistic model is an approach that has been
tried
        <xref ref-type="bibr" rid="ref10">(De Belder et al., 2010)</xref>
        . Probabilistic models
are obtained from lexical simplifications, which
have previously done applying E2R guidelines, as
in the Simple Wikipedia. McCarthy and Navigli
        <xref ref-type="bibr" rid="ref26 ref31">(McCarthy and Navigli, 2007)</xref>
        introduce work to
propose candidates to replace a word using
contexts. In Semeval 2012, English Lexical
Simplification challenge
        <xref ref-type="bibr" rid="ref40">(Specia et al., 2012)</xref>
        with ten
participant systems, the evaluation results showed
that proposals based on frequency give good
results comparing to other sophisticated systems.
      </p>
      <p>
        Focusing on research devoted to synonym
substitution in Spanish texts, lack of semantic
resources is a handicap. A recent work is described
in
        <xref ref-type="bibr" rid="ref5">(Bott et al., 2012)</xref>
        , LexSiS system that uses
Spanish OpenThesaurus to build a vector space
model according to the distributional hypothesis
that establishes that different uses of a word tend
to appear in different lexical contexts. A vector is
built in a window of nine words around each
wordsense in a corpus extracted from the
OpenThesarus and compared using the cosine similarity
combined with word frequency and word length.
This approach can be enhanced including
rulebased lexical simplification, see
        <xref ref-type="bibr" rid="ref13 ref5">(Drndarevic et
al., 2012)</xref>
        , where some patterns that avoid
incorrect substitutions are defined, for instance, to
replace reporting verbs (confirm, suggest, explain,
etc.) that leaves correct syntactic structures as
well as other editing transformations (numerical
expressions or periphrasis). Following the same
approach, CASSA method is reported in
(BaezaYates et al., 2015) where the Spanish corpus used
to extract word occurrences is the Google Books
Ngram corpus that contains real web
frequencies. This work also obtains word senses from
OpenThesaurus.
      </p>
      <p>
        But before simplifying we have to know the
level of readability and understandability of a text
by using complexity measures. There are simple
measures based on frequency of words in texts as
well as length of phrases, FOX index
        <xref ref-type="bibr" rid="ref18">(Gunning,
1986)</xref>
        , Flesch-Kinaid
        <xref ref-type="bibr" rid="ref22">(Kincaid et al., 1975)</xref>
        measures are used in English. In Spanish texts, several
indexes have been proposed to measure the
structural complexity of a text
        <xref ref-type="bibr" rid="ref1">(Anula, 2007)</xref>
        : the
number or verbal predicates in subordinate clauses,
and the index of sentence recursion (a measure that
counts the number of nested clauses in the text).
To measure the lexical complexity two indexes are
proposed: an index of low frequency words (the
number of content words1 with low frequency
divided by the total number of lexical words) and an
index of lexical density (number of distinct
content words /total of discourse segments2). Finally,
other indexes such as the average length of
sentences and average length of words (syllables)
although they are criticized. These indexes have to
be validated by the end users. Knowing the
readability level of a document, users have the
opportunity to choose the most suitable text, from a
collection of documents delivering the same
information
        <xref ref-type="bibr" rid="ref13 ref29 ref35 ref40 ref41 ref5">(Sbattella and Tedesco, 2012)</xref>
        .
      </p>
      <p>With respect to Spanish corpora for extraction
of frequencies and word contexts, the CREA3
corpus available online is not a useful resource when
domain specific texts are required (for instance,
1A content word is a word with meaning (nouns, verbs,
adjectives and adverbs)
2sentences or phrases
3http://corpus.rae.es/creanet.html
biology or chemical texts). The latest version of
June 2008 contains one hundred and sixty million
of documents (from journals, books and
newspapers covering more than one hundred subjects). In
2018 The Royal Spanish Academy (RAE) will
deliver the CORPES XXI, a higher Spanish corpus
with four hundred million of forms.</p>
      <p>
        Finally, there are specific works to simplify
numerical expressions. Bautista and Saggion (2014)
        <xref ref-type="bibr" rid="ref30 ref4">(Bautista and Saggion, 2014)</xref>
        propose a rule-based
lexical component that simplifies numerical
expressions in Spanish texts. This work makes
news articles more accessible to certain readers by
rewriting difficult numerical expressions in a
simpler way.
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>The EasyLecto system</title>
      <p>The EasyLecto system aims to simplify the drug
package leaflets, in particular, replacing the terms
describing adverse drug reactions with synonyms
that are easier to understand for the patients.</p>
      <p>Figure 1 illustrates the EasyLecto system
architecture. The first module of the EasyLecto
system aims to automatically annotate adverse
drug reactions in texts. This module uses a
dictionary-based approach that combines
terminological resources, such as MedDRA, the ATC
system (a drug classification system developed
by the World Health Organization) or CIMA (a
database of medicines approved in Spain), or
dictionaries gathered from websites about health and
medicines such as MedLinePlus4, vademecum.es5
or prospectos.net6. The reader can find a
detailed description of the NER module in
(SeguraBedmar et al., 2015).</p>
      <p>Once adverse drug reactions are automatically
identified in texts, a set of synonyms is proposed
for each one of them. MedDRA7 is a medical
terminology dictionary about events associated with
drugs. It is a multilingual dictionary (11
languages) and its main goal is to provide a
classification system for efficient communication of adverse
drug reactions data between countries. MedDRA
is composed of a five-level hierarchy. The most
specific level, ”Lowest Level Terms” (LLTs),
contains a total of 72,072 terms that express how
information is communicated in practice. The main
4https://www.nlm.nih.gov/medlineplus/spanish/
5http://www.vademecum.es
6https://www.prospectos.net
7http://www.meddra.org/
advantage of MedDRA is that its structured format
allows easily obtaining a list of possible adverse
drug reactions and their synonyms. Thus, we
decided to use MedDRA as a source of synonyms for
adverse drug reactions. Moreover, for a given
effect in MedDRA, we used its longest synonym as
definition for the effect.</p>
      <p>
        The following step is to select the appropriate
synonym, that is, the simplest synonym. The more
common a term is in a collection of texts, the more
familiar the term is likely to be to the reader
        <xref ref-type="bibr" rid="ref15">(Elhadad, 2006)</xref>
        . Thus, our system proposes those
synonyms with higher frequency. In order to know
how common a word is, we gathered a large
collection of texts such as the MedLinePlus articles 8,
and indexed it in order to obtain the frequency of
each drug effect.
      </p>
      <p>MedLinePlus is an online resource with health
information for patients, which contains more than
1,000 articles about diseases and 6,000 articles
about medicines. The Spanish version is one of
the most comprehensive and trusted Spanish
language health websites at the moment. We
developed a web crawler to browse and download pages
related to drugs and diseases from the
MedLinePlus website. Each MedLinePlus article provides
exhaustive information about a given medical
concept, and also proposes a list of related health
topics, which can be considered as synonyms of this
concept. Moreover, an article related to a given
medical concept can also be used to obtain the
definition of this concept by getting its first
sentence. Finally, all downloaded articles, the
definitions (first sentence of each article) and their
related health topics were translated into JSON
objects in order to create an index (see Figure 2)
using ElasticSearch9, an open source search engine.</p>
      <p>All told, the EasyLecto system proposes a
definition and a set of synonyms from MedDRA, as
well as a definition and a set of synonyms from
MedLinePlus, for each drug effect. Then, the
frequency of each synonym is calculated using the
index built from MedLinePlus, and finally the
synonym with the highest frequency is selected as the
simplest synonym.</p>
      <p>Due to the horizontal scalability provided by
ElasticSearch, it is possible to index large
collections of documents, as is the case of the
MedlinePlus. The main advantage of ElasticSearch
is its capacity to create distributed systems by
specifying only the configuration of the hierarchy
of nodes. Then, ElasticSearch is self-managed
to maintain better fault tolerance and load
distribution. Another important advantage of
ElasticSearch is that it does not require very high
computing power and a high storage capacity to index
large collections. In this study, ElasticSearch
(version 2.2) was installed on a Ubuntu Server 14.04
8https://www.nlm.nih.gov/medlineplus/spanish/
9http://elasticsearch.org
with 8GB of RAM and 500GB of disk space.</p>
      <p>A demo of the EasyLecto system is available at:
http://jacky.uc3m.es/EasyLecto/. This tool allows
to load a document highlighting the adverse drug
reactions (in blue) (see Figure 3). If the user
selects any of these adverse drug reactions, the tool
displays a popup window with information about
the definitions and synonyms proposed by the
system. Figure 4 shows the synonyms and
definitions proposed for the effect ’dispepsia’
(dyspepsia). While the most frequent MedDRA synonym
was ’indigestio´n’ (indigestion), the most common
synonym from MedLinePlus was ’enfermedades
del esto´mago’ (stomach diseases).
4</p>
    </sec>
    <sec id="sec-4">
      <title>Evaluation</title>
      <p>
        The dataset used for the evaluation is the
EasyDPL (easy drug package leaflets) corpus10, which
contains 306 package leaflets annotated with 1,400
adverse drug reactions and their simplest
synonyms. The corpus was manually annotated by
three trained annotators. The quality and
consistency of the corpus were evaluated by measuring
inter-annotator agreement (IAA). IAA also
determines the complexity of the task and provides an
upper bound on the performance of the automatic
systems for the simplification of adverse drug
reactions in drug package leaflets. In particular, the
Fleiss’ kappa
        <xref ref-type="bibr" rid="ref17">(Fleiss, 1971)</xref>
        was calculated, which
is an extension of Cohen’s kappa
        <xref ref-type="bibr" rid="ref7">(Cohen, 1960)</xref>
        that measures the degree of consistency for two or
more annotators. The assessment showed a kappa
of 0.709, which is considered substantial on the
Landis and Koch scale
        <xref ref-type="bibr" rid="ref23">(Landis and Koch, 1977)</xref>
        .
      </p>
      <p>For each drug effect annotated in the EasyDDI
corpus, the evaluation consisted in comparing the
gold-standard synonym, that is, the synonym
proposed by the human annotators, to the simplest
synonym, that is, the synonym with the highest
frequency in the index built from the
MedLinePlus articles. Since we used two different
resources, MedDRa and MedLinePlus, in order to
achieve the set of synonym candidates, we
evaluated the simplest synonym from each of the
resources. Thus, for the synonym obtained from
MedLinePlus, EasyLecto achieves an accuracy of
68.7%, while for the MedDRA synonym, the
accuracy is much lower (around 37.2%). This is
mainly due to MedDRA being a highly specific
standardized medical terminology, which implies
its terms are not familiar to most people.
MedLinePlus on the other hand is a health information
website for patients, which uses a more readable
language and a lay vocabulary.</p>
      <p>We conducted an error analysis in order to
obtain the main causes of false positives and false
negatives in our system. In particular, we studied
in detail a random sample of 30 documents.
Table 1 presents some errors that our system makes
on the EasyDPL corpus. Most errors are due
to the absence of a simpler synonym for a term;
some terms could only be explained by a small
sentence or phrase (for example, terms such as
akathisia or eosinophilia). Another cause of
error was that some terms were replaced by their
hypernyms in the gold-standard corpus (for
example, allergic alveolitis was substituted by allergy),
whereas the system failed because it does not
exploit the hierarchical relationships between terms
and is not able to propose more general terms as
synonyms for a specific term. Some errors, such
as dysphoria-hoarseness or diaphoresis -
sweating, may occur due to the lack of synonyms in
the resources. An approach based on a word
vector model able to compute the similarity between
words based on their contexts, could reduce such
errors.</p>
      <p>In addition to the quantitative evaluation, we
also used SurveyMonkey to collect some quick
user feedback on the EasyLecto system 11. We
defined a survey with 10 closed-ended questions,
in which users should pick just one answer from
a list of given options. We asked users about the
usefulness and the performance of the EasyLecto,
as well as about its usability, design and visual
appeal. A total of 26 users completed the survey,
10http://labda.inf.uc3m.es/doku.php?id=en:labda recursosPLN
11https://es.surveymonkey.com/r/8HMVJKV
most of them being software engineers or PhD
students in computer science. The analysis of the
survey shows that most users have positive opinions
about the EasyLecto system. Almost 97% of users
think that the EasyLecto system helps to simplify
drug package leaflets. Regarding the definitions
proposed by the system, 75% of users believe that
the definitions help to understand the text. Almost
30% of them would like to obtain three or more
synonyms from the system. Around 81% of users
think that the EasyLecto has a friendly interface.
5</p>
    </sec>
    <sec id="sec-5">
      <title>Conclusions and future work</title>
      <p>
        Although drug package leaflets should be
designed and written ensuring complete
understanding of their contents, several factors can have an
influence on patient understanding of drug
package leaflets. Low literacy is directly associated
with limited understanding and misinterpretation
of these documents
        <xref ref-type="bibr" rid="ref8 ref8 ref9 ref9">(Davis et al., 2006b; Davis et
al., 2006a)</xref>
        . Older people are more likely to have
lower literacy skills, as well as decreased memory
and poorer reading comprehension (Kutner et al.,
2006). Therefore, low literacy along with older
age may lead to an unintentional non-compliance
or inappropriate use of drugs, leading to
dangerous consequences for patients, such as
therapeutic failure or adverse drug reactions. Several
studies
        <xref ref-type="bibr" rid="ref25 ref32 ref33">(March et al., 2010; Pires et al., 2015;
Pin˜eroLo´pez et al., 2016)</xref>
        have shown that there is an
urgent need to improve the quality of drug package
leaflets because they are usually too difficult to
understand for patients, and this could be a potential
source of drug related problems, such as
medication errors and adverse drug reactions. In
particular, patients have problems to understand those
sections describing dosages and adverse drug
reactions.
      </p>
      <p>The EasyLecto system aims the simplification
of drug package leaflets, in particular, the
simplification of terms describing adverse drug
reactions by synonyms that are easier to understand by
patients. The system uses a dictionary-based
approach in order to automatically identify adverse
drug reactions in drug package leaflets. MedDRA
and MedLinePlus are used as sources of synonyms
and definitions for these effects. Our main
hypothesis is that a simple word will likely be more
common in a collection of texts than their more
difficult synonyms. We built an index from a large
collection of texts such as MedLinePlus. This
index provides us information about how common
a word is. EasyLecto was evaluated on a
goldstandard corpus with 306 texts manually annotated
by three trained experts. Experiments show an
accuracy of 68.7% for the MedLinePlus synonym
and 37.1% for the MedDRA synonym. Therefore,
resources that have been specially written for
patients are a better source of simpler synonyms that
the specialized terminological resources (such as
MedDRA). On the other hand, the error analysis
shows that some of the system answers might as
well be valid and simple synonyms, even though
they are not the same as proposed by the
goldstandard corpus. In order to obtain a more realistic
evaluation, we plan to extend the EasyDPL corpus
by adding several simpler synonyms for each term.</p>
      <p>In addition to the quantitative evaluation, the
subjective impression of 26 users was documented
by a simple questionnaire published in
SurveyMonkey. In general, users have positive
perceptions of the EasyLecto system. We are aware that
our evaluation system based on user experience
has a lot of shortcomings (e.g., the number of users
is very small and they are not representative of the
general public). Therefore, we plan to extend and
improve the evaluation with a large set of users
that includes elderly users, people with disabilities
or with low literacy levels.</p>
      <p>
        In this work, we only focus on the
simplification of adverse drug reactions, however we plan to
extend our approach in order to simplify not only
other medical concepts (such as diseases, medical
procedures, medical tests, etc), but also complex
words from open-domain texts. As future work,
we also plan to integrate additional resources such
as BabelNet
        <xref ref-type="bibr" rid="ref13 ref29 ref35 ref40 ref41 ref5">(Navigli and Ponzetto, 2012)</xref>
        or the
UMLS Metathesaurus
        <xref ref-type="bibr" rid="ref24">(Lindberg et al., 1993)</xref>
        . In
addition to providing broader coverage for terms
and more synonyms, these resources will allow to
develop a multilingual simplification system.
      </p>
      <p>
        To the best of our knowledge, while word vector
models based on n-grams have already been used
        <xref ref-type="bibr" rid="ref5">(Bott et al., 2012)</xref>
        , word vector models trained
using deep learning techniques have not been
explored for the task of simplification yet. We also
plan to study the use of word embeddings learned
by Word2Vec
        <xref ref-type="bibr" rid="ref27">(Mikolov et al., 2013)</xref>
        or Glove
        <xref ref-type="bibr" rid="ref30">(Pennington et al., 2014)</xref>
        . One important
advantage of these models is that they allow to
compute the similarity between terms without the need
of using synonym dictionaries that are generally
      </p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This work was supported by
eGovernAbilityAccess project (TIN2014-52665-C2-2-R).
index, fog count and flesch reading ease formula)
for navy enlisted personnel. Technical report, DTIC
Document.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <given-names>Alberto</given-names>
            <surname>Anula</surname>
          </string-name>
          .
          <year>2007</year>
          .
          <article-title>Tipos de textos, complejidad lingu¨´ıstica y facilitacio´n de la lectura</article-title>
          .
          <source>In Actas del IV Congreso de la Asociacio´</source>
          n Asia´tica de Hispanistas.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <given-names>Ricardo</given-names>
            <surname>Baeza-Yates</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Luz</given-names>
            <surname>Rello</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Julia</given-names>
            <surname>Dembowski</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Cassa: A context-aware synonym simplification algorithm</article-title>
          .
          <source>In Human Language Technologies: The 2015 Annual Conference of the North American Chapter of the ACL, page 13801385.</source>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <given-names>Thimoty</given-names>
            <surname>Barbieri</surname>
          </string-name>
          ,
          <string-name>
            <surname>Antonio</surname>
            <given-names>BIANCHI</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Licia</surname>
            <given-names>SBATTELLA</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ferdinando</surname>
            <given-names>CARELLA</given-names>
          </string-name>
          , and Marco FERRA.
          <year>2005</year>
          .
          <article-title>Multiabile: A multimodal learning environment for the inclusion of impaired e-learners using tactile feedbacks, voice, gesturing, and text simplification</article-title>
          .
          <source>Assistive Technology: From Virtuality to Reality</source>
          ,
          <volume>16</volume>
          (
          <issue>1</issue>
          ):
          <fpage>406</fpage>
          -
          <lpage>410</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <given-names>Susana</given-names>
            <surname>Bautista</surname>
          </string-name>
          and
          <string-name>
            <given-names>Horacio</given-names>
            <surname>Saggion</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Can numerical expressions be simpler? implementation and demostration of a numerical simplification system for spanish</article-title>
          .
          <source>In LREC</source>
          , pages
          <fpage>956</fpage>
          -
          <lpage>962</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <given-names>Stefan</given-names>
            <surname>Bott</surname>
          </string-name>
          , Luz Rello, Biljana Drndarevic, and
          <string-name>
            <given-names>Horacio</given-names>
            <surname>Saggion</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>Can spanish be simpler? lexsis: Lexical simplification for spanish</article-title>
          .
          <source>In Proceedings of COLING 2012</source>
          , pages
          <fpage>357</fpage>
          -
          <lpage>374</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <given-names>Raman</given-names>
            <surname>Chandrasekar</surname>
          </string-name>
          , Christine Doran, and
          <string-name>
            <given-names>Bangalore</given-names>
            <surname>Srinivas</surname>
          </string-name>
          .
          <year>1996</year>
          .
          <article-title>Motivations and methods for text simplification</article-title>
          .
          <source>In Proceedings of the 16th conference on Computational linguistics-Volume</source>
          <volume>2</volume>
          , pages
          <fpage>1041</fpage>
          -
          <lpage>1044</lpage>
          . Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <given-names>Jacob</given-names>
            <surname>Cohen</surname>
          </string-name>
          .
          <year>1960</year>
          .
          <article-title>A coefficient of agreement for nominal scale</article-title>
          .
          <source>Educ Psychol Meas</source>
          ,
          <volume>20</volume>
          :
          <fpage>37</fpage>
          -
          <lpage>46</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Terry C Davis</surname>
            , Michael S Wolf, Pat F Bass,
            <given-names>Mark</given-names>
          </string-name>
          <string-name>
            <surname>Middlebrooks</surname>
          </string-name>
          , Estela Kennen, David W Baker, Charles L Bennett,
          <string-name>
            <surname>Ramon</surname>
            Durazo-Arvizu, Anna Bocchini,
            <given-names>Stephanie</given-names>
          </string-name>
          <string-name>
            <surname>Savory</surname>
          </string-name>
          , et al. 2006a.
          <article-title>Low literacy impairs comprehension of prescription drug warning labels</article-title>
          .
          <source>Journal of general internal medicine</source>
          ,
          <volume>21</volume>
          (
          <issue>8</issue>
          ):
          <fpage>847</fpage>
          -
          <lpage>851</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Terry C Davis</surname>
          </string-name>
          , Michael S Wolf, Pat F Bass,
          <article-title>Jason A Thompson, Hugh H Tilson, Marolee Neuberger, and Ruth M Parker. 2006b. Literacy and misunderstanding prescription drug labels</article-title>
          .
          <source>Annals of Internal Medicine</source>
          ,
          <volume>145</volume>
          (
          <issue>12</issue>
          ):
          <fpage>887</fpage>
          -
          <lpage>894</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <string-name>
            <surname>Jan De Belder</surname>
          </string-name>
          , Koen Deschacht, and
          <string-name>
            <surname>Marie-Francine Moens</surname>
          </string-name>
          .
          <year>2010</year>
          .
          <article-title>Lexical simplification</article-title>
          .
          <source>In Proceedings of ITEC2010: 1st international conference on interdisciplinary research on technology, education and communication.</source>
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <given-names>Siobhan</given-names>
            <surname>Lucy Devlin</surname>
          </string-name>
          .
          <year>1999</year>
          .
          <article-title>Simplifying natural language for aphasic readers</article-title>
          .
          <source>Ph.D. thesis</source>
          , University of Sunderland.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <given-names>Bonnie</given-names>
            <surname>Dorr</surname>
          </string-name>
          , David Zajic, and Richard Schwartz.
          <year>2003</year>
          .
          <article-title>Hedge trimmer: A parse-and-trim approach to headline generation</article-title>
          .
          <source>In Proceedings of the HLTNAACL 03 on Text summarization workshop-</source>
          Volume
          <volume>5</volume>
          , pages
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          . Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <given-names>Biljana</given-names>
            <surname>Drndarevic</surname>
          </string-name>
          , Sanja Sˇ tajner, and
          <string-name>
            <given-names>Horacio</given-names>
            <surname>Saggion</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>Reporting simply: A lexical simplification strategy for enhancing text accessibility</article-title>
          .
          <source>In Proceedings of Easy-to-Read on the Web Symposium.</source>
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <surname>EC.</surname>
          </string-name>
          <year>2009</year>
          .
          <article-title>Guideline on the readability of the labelling and package leaflet of medicinal products for human use</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <source>Noe´mie Elhadad</source>
          .
          <year>2006</year>
          .
          <article-title>Comprehending technical texts: predicting and defining unfamiliar terms</article-title>
          .
          <source>In AMIA.</source>
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <given-names>Council</given-names>
            <surname>EU</surname>
          </string-name>
          .
          <year>2001</year>
          .
          <article-title>Directive 2001/83/ec of the european parliament and of the council of 6 november 2001 on the community code relating to medicinal products for human use</article-title>
          .
          <source>Official Journal L</source>
          ,
          <volume>311</volume>
          (
          <issue>28</issue>
          ):
          <fpage>11</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <surname>Joseph L Fleiss</surname>
          </string-name>
          .
          <year>1971</year>
          .
          <article-title>Measuring nominal scale agreement among many raters</article-title>
          .
          <source>Psychological bulletin</source>
          ,
          <volume>76</volume>
          (
          <issue>5</issue>
          ):
          <fpage>378</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <given-names>Robert</given-names>
            <surname>Gunning</surname>
          </string-name>
          .
          <year>1986</year>
          .
          <article-title>The technique of clear writing</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          <string-name>
            <given-names>Siddhartha</given-names>
            <surname>Jonnalagadda</surname>
          </string-name>
          and
          <string-name>
            <given-names>Graciela</given-names>
            <surname>Gonzalez</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>Biosimplify: an open source sentence simplification engine to improve recall in automatic biomedical information extraction</article-title>
          .
          <source>arXiv preprint arXiv:1107</source>
          .
          <fpage>5744</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          <string-name>
            <given-names>Siddhartha</given-names>
            <surname>Jonnalagadda</surname>
          </string-name>
          , Luis Tari, Jo¨rg Hakenberg, Chitta Baral, and
          <string-name>
            <given-names>Graciela</given-names>
            <surname>Gonzalez</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <article-title>Towards effective sentence simplification for automatic processing of biomedical text</article-title>
          .
          <source>In Proceedings of Human Language Technologies</source>
          :
          <article-title>The 2009 Annual Conference of the North American Chapter of the Association for Computational Linguistics</article-title>
          , Companion Volume:
          <source>Short Papers</source>
          , pages
          <fpage>177</fpage>
          -
          <lpage>180</lpage>
          . Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          <string-name>
            <given-names>Sasikiran</given-names>
            <surname>Kandula</surname>
          </string-name>
          , Dorothy Curtis, and Qing ZengTreitler.
          <year>2010</year>
          .
          <article-title>A semantic and syntactic text simplification tool for health content</article-title>
          .
          <source>In AMIA Annu Symp Proc</source>
          , volume
          <volume>2010</volume>
          , pages
          <fpage>366</fpage>
          -
          <lpage>70</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          <string-name>
            <surname>J Peter Kincaid</surname>
          </string-name>
          , Robert P Fishburne Jr, Richard L Rogers, and Brad S Chissom.
          <year>1975</year>
          .
          <article-title>Derivation of new readability formulas (automated readability Mark Kutner, Elizabeth Greenberg</article-title>
          , and
          <string-name>
            <given-names>Justin</given-names>
            <surname>Baer</surname>
          </string-name>
          .
          <year>2006</year>
          .
          <article-title>A first look at the literacy of america's adults in the 21st century</article-title>
          .
          <source>nces</source>
          <year>2006</year>
          -
          <volume>470</volume>
          . National Center for Education Statistics.
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          <string-name>
            <given-names>J Richard</given-names>
            <surname>Landis</surname>
          </string-name>
          and Gary G Koch.
          <year>1977</year>
          .
          <article-title>The measurement of observer agreement for categorical data</article-title>
          .
          <source>biometrics</source>
          , pages
          <fpage>159</fpage>
          -
          <lpage>174</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          <string-name>
            <given-names>Donald A Lindberg</given-names>
            , Betsy L Humphreys, and
            <surname>Alexa T McCray</surname>
          </string-name>
          .
          <year>1993</year>
          .
          <article-title>The unified medical language system</article-title>
          .
          <source>Methods of information in medicine</source>
          ,
          <volume>32</volume>
          (
          <issue>4</issue>
          ):
          <fpage>281</fpage>
          -
          <lpage>291</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          <string-name>
            <surname>Cerda´ JC March</surname>
          </string-name>
          , Rodr´ıguez MA Prieto,
          <article-title>Azarola A Ruiz, Lorda P Simo´n, Cantalejo I Barrio,</article-title>
          and
          <string-name>
            <given-names>Alina</given-names>
            <surname>Danet</surname>
          </string-name>
          .
          <year>2010</year>
          .
          <article-title>[quality improvement of health information included in drug information leaflets. patient and health professional expectations]</article-title>
          .
          <source>Atencio´</source>
          n primaria/Sociedad Espan˜ola de Medicina de Familia y Comunitaria,
          <volume>42</volume>
          (
          <issue>1</issue>
          ):
          <fpage>22</fpage>
          -
          <lpage>27</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          <string-name>
            <given-names>Diana</given-names>
            <surname>McCarthy</surname>
          </string-name>
          and
          <string-name>
            <given-names>Roberto</given-names>
            <surname>Navigli</surname>
          </string-name>
          .
          <year>2007</year>
          .
          <article-title>Semeval2007 task 10: English lexical substitution task</article-title>
          .
          <source>In Proceedings of the 4th International Workshop on Semantic Evaluations</source>
          , pages
          <fpage>48</fpage>
          -
          <lpage>53</lpage>
          . Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          <string-name>
            <given-names>Tomas</given-names>
            <surname>Mikolov</surname>
          </string-name>
          , Ilya Sutskever, Kai Chen, Greg S Corrado, and
          <string-name>
            <given-names>Jeff</given-names>
            <surname>Dean</surname>
          </string-name>
          .
          <year>2013</year>
          .
          <article-title>Distributed representations of words and phrases and their compositionality</article-title>
          .
          <source>In Advances in neural information processing systems</source>
          , pages
          <fpage>3111</fpage>
          -
          <lpage>3119</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          <string-name>
            <given-names>George A</given-names>
            <surname>Miller</surname>
          </string-name>
          .
          <year>1995</year>
          .
          <article-title>Wordnet: a lexical database for english</article-title>
          .
          <source>Communications of the ACM</source>
          ,
          <volume>38</volume>
          (
          <issue>11</issue>
          ):
          <fpage>39</fpage>
          -
          <lpage>41</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          <string-name>
            <given-names>Roberto</given-names>
            <surname>Navigli</surname>
          </string-name>
          and Simone Paolo Ponzetto.
          <year>2012</year>
          .
          <article-title>Babelnet: The automatic construction, evaluation and application of a wide-coverage multilingual semantic network</article-title>
          .
          <source>Artificial Intelligence</source>
          ,
          <volume>193</volume>
          :
          <fpage>217</fpage>
          -
          <lpage>250</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          <string-name>
            <surname>Jeffrey</surname>
            <given-names>Pennington</given-names>
          </string-name>
          , Richard Socher, and
          <string-name>
            <given-names>Christopher D</given-names>
            <surname>Manning</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>Glove: Global vectors for word representation</article-title>
          .
          <source>In EMNLP</source>
          , volume
          <volume>14</volume>
          , pages
          <fpage>1532</fpage>
          -
          <lpage>1543</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          <string-name>
            <given-names>Sarah E</given-names>
            <surname>Petersen</surname>
          </string-name>
          and
          <string-name>
            <given-names>Mari</given-names>
            <surname>Ostendorf</surname>
          </string-name>
          .
          <year>2007</year>
          .
          <article-title>Text simplification for language learners: a corpus analysis</article-title>
          .
          <source>In Proceedings of Workshop on Speech and Language Technology for Education</source>
          , pages
          <fpage>69</fpage>
          -
          <lpage>72</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          <article-title>A´ ngeles Mar´ıa Pin˜ero-Lo´pez</article-title>
          , Pilar Modamio,
          <string-name>
            <given-names>F.</given-names>
            <surname>Cecilia Lastra</surname>
          </string-name>
          , and L. Eduardo Marin˜o.
          <year>2016</year>
          .
          <article-title>Readability analysis of the package leaflets for biological medicines available on the internet between 2007 and 2013: An analytical longitudinal study</article-title>
          .
          <source>J Med Internet Res</source>
          ,
          <volume>18</volume>
          (
          <issue>5</issue>
          ):
          <fpage>e100</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          <string-name>
            <given-names>Carla</given-names>
            <surname>Pires</surname>
          </string-name>
          , Marina Viga´rio, and
          <string-name>
            <given-names>Afonso</given-names>
            <surname>Cavaco</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Readability of medicinal package leaflets: a systematic review</article-title>
          . Revista de saude publica,
          <volume>49</volume>
          :
          <fpage>1</fpage>
          -
          <lpage>13</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          <string-name>
            <given-names>Horacio</given-names>
            <surname>Saggion</surname>
          </string-name>
          , Elena Go´mez Mart´ınez, Esteban Etayo, Alberto Anula, and
          <string-name>
            <given-names>Lorena</given-names>
            <surname>Bourg</surname>
          </string-name>
          .
          <year>2011</year>
          .
          <article-title>Text simplification in simplext. making text more accessible</article-title>
          .
          <source>Procesamiento del lenguaje natural</source>
          ,
          <volume>47</volume>
          :
          <fpage>341</fpage>
          -
          <lpage>342</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          <string-name>
            <given-names>Licia</given-names>
            <surname>Sbattella</surname>
          </string-name>
          and
          <string-name>
            <given-names>Roberto</given-names>
            <surname>Tedesco</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>Calculating text complexity during the authoring phase</article-title>
          .
          <source>In Proceedings of Easy-to-Read on the Web Symposium.</source>
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          <string-name>
            <given-names>Isabel</given-names>
            <surname>Segura-Bedmar</surname>
          </string-name>
          , Paloma Mart´ınez, Ricardo Revert, and Julia´n Moreno-Schneider.
          <year>2015</year>
          .
          <article-title>Exploring spanish health social media for detecting drug effects. BMC medical informatics and decision making, 15(2):1</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          <string-name>
            <given-names>Matthew</given-names>
            <surname>Shardlow</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>A survey of automated text simplification</article-title>
          .
          <source>International Journal of Advanced Computer Science and Applications</source>
          ,
          <volume>4</volume>
          (
          <issue>1</issue>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          <string-name>
            <given-names>Advaith</given-names>
            <surname>Siddharthan</surname>
          </string-name>
          .
          <year>2002</year>
          .
          <article-title>Resolving attachment and clause boundary ambiguities for simplifying relative clause constructs</article-title>
          .
          <source>In Proceedings of the Student Workshop, 40th Meeting of the Association for Computational Linguistics (ACL02)</source>
          , pages
          <fpage>60</fpage>
          -
          <lpage>65</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          <string-name>
            <given-names>Advaith</given-names>
            <surname>Siddharthan</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <article-title>A survey of research on text simplification</article-title>
          .
          <source>ITL-International Journal of Applied Linguistics</source>
          ,
          <volume>165</volume>
          (
          <issue>2</issue>
          ):
          <fpage>259</fpage>
          -
          <lpage>298</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          <string-name>
            <given-names>Lucia</given-names>
            <surname>Specia</surname>
          </string-name>
          , Sujay Kumar Jauhar, and
          <string-name>
            <given-names>Rada</given-names>
            <surname>Mihalcea</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>Semeval-2012 task 1: English lexical simplification</article-title>
          .
          <source>In Proceedings of the First Joint Conference on Lexical and Computational SemanticsVolume 1: Proceedings of the main conference and the shared task, and Volume 2: Proceedings of the Sixth International Workshop on Semantic Evaluation</source>
          , pages
          <fpage>347</fpage>
          -
          <lpage>355</lpage>
          . Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref41">
        <mixed-citation>
          <string-name>
            <given-names>Irina</given-names>
            <surname>Temnikova</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>Text Complexity and Text Simplification in the Crisis Management domain</article-title>
          .
          <source>Ph.D. thesis</source>
          , University of Wolverhampton.
        </mixed-citation>
      </ref>
      <ref id="ref42">
        <mixed-citation>
          <string-name>
            <given-names>Zhemin</given-names>
            <surname>Zhu</surname>
          </string-name>
          , Delphine Bernhard, and
          <string-name>
            <given-names>Iryna</given-names>
            <surname>Gurevych</surname>
          </string-name>
          .
          <year>2010</year>
          .
          <article-title>A monolingual tree-based translation model for sentence simplification</article-title>
          .
          <source>In Proceedings of the 23rd international conference on computational linguistics</source>
          , pages
          <fpage>1353</fpage>
          -
          <lpage>1361</lpage>
          . Association for Computational Linguistics.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>