<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Contextual Definition Generation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jeffrey T. Yarbro</string-name>
          <email>jyarbro2@memphis.edu</email>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andrew M. Olney</string-name>
          <email>aolney@memphis.edu</email>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Memphis</institution>
          ,
          <addr-line>Memphis, TN 38152</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper explores the concept of dynamically generating definitions using a deep-learning model. We do this by creating a dataset that contains definition entries and contexts associated with each definition. We then fine-tune a GPT-2 based model on the dataset to allow the model to generate contextual definitions. We evaluate our model with human raters by generating definitions using two context types: short-form (the word used in a sentence) and long-form (the word used in a sentence along with the prior and following sentences). Results indicate that the model performed significantly better when generating definitions using short-form contexts. Additionally, we evaluate our model against human-generated definitions. The results show promise for the model, showing that the model was able to match human-level fluency. However, while it was able to reach human-level accuracy in some instances, it failed in others.</p>
      </abstract>
      <kwd-group>
        <kwd>GPT-2</kwd>
        <kwd>contextual definitions</kwd>
        <kwd>human evaluation</kwd>
        <kwd>definition generation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Prior studies have suggested that 95% of words in a text must be within a reader's
vocabulary for adequate reading comprehension to occur [10, 15]. This presents a problem
for academic texts, which often use low-frequency words to describe key concepts
related to their field. For readers to comprehend these concepts, they must first acquire
the field-specific vocabulary. One of the classical ways to help with this problem is to
have a glossary near the end of the textbook or a list of key terms at the beginning of
the chapter, allowing the reader to have a relatively easy way to familiarize themselves
with words that they are unfamiliar with and reference these definitions during reading.
This classical approach in paper-based textbooks has a few notable issues: (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) the reader
is forced to change pages and find the term that they are looking for (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) the number of
definitions is limited to a small subset of words within the book (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ) the definition might
not be appropriate to the context, as happens when a word has multiple definitions.
      </p>
      <p>
        Intelligent systems have made the process of finding a definition much easier. Rather
than turning pages or firin1g up a search engine, readers can now click on a piece of
text, and a pop-up will appear displaying the definition for the selected term [5, 7].
These definitions are typically acquired by querying a definition database. While this
is a good solution, it does come with some issues. (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) The source of definitions must
1 Copyright © 2021 for this paper by its authors. Use permitted under Creative
Commons License Attribution 4.0 International (CC BY 4.0).
contain the definition for the word in the context displayed. (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) If the word has multiple
definitions, one must either display a list of definitions or deploy a word sense
disambiguation model to find the most fitting definition. (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ) Definitions are not tuned to the
precise context of the word.
      </p>
      <p>
        This paper explores the concept of using a deep-learning model to generate dynamic,
contextual definitions by paying attention to the surrounding context of the word. We
do this by creating a new dataset consisting of words, a definition for each word, and a
list of contexts associated with each definition. We then fine-tune a GPT-2 based model
on this dataset, resulting in a model capable of autoregressively generating a definition
for any English word with only the word and a context as inputs. We assess the model
using human evaluation with three research questions in mind: (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) What type of context
provides the best initialization for the model (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) How does the model perform relative
to human-generated definitions. (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ) Is the model biased towards any particular subject?
      </p>
      <p>We answer these questions by analyzing how human raters rate machine-generated
and human-generated definitions for terms from 5 college-level textbooks in terms of
accuracy and grammatical fluency.</p>
    </sec>
    <sec id="sec-2">
      <title>Data Collection and Training</title>
      <sec id="sec-2-1">
        <title>Data Collection</title>
        <p>A key constraint before beginning data collection was to find sources where word
entries had both a definition and a context that matched that definition. A matching
definition and contextual pair are important since sampling random sentences/paragraphs
the term is within and attempting to find the most appropriate definition could lead to
errors. We did not impose strict types of contexts. Contexts could be words in example
sentences, words in example paragraphs, hypernyms, synonyms, hyponyms, etc. The
primary goal was to give the model enough of a contextual clue to internally
disambiguate the sense of the word and generate a definition that fit that sense.</p>
        <p>
          With this in mind, we collected training data from the following sources: (
          <xref ref-type="bibr" rid="ref1">1</xref>
          )
Wiktionary: Extracted definition/context data by cleaning the March 2021 version of the
XML dump file [23]. Contexts contained synonyms, example sentences/paragraphs,
hypernyms, hyponyms, and sense tags. (
          <xref ref-type="bibr" rid="ref2">2</xref>
          ) Lexico: Scraped Lexico for definitions,
example sentences, synonyms, and sense tags [11] (
          <xref ref-type="bibr" rid="ref3">3</xref>
          ) WordNet: Used the NLTK
implementation of WordNet to acquire gloss entries and sense information [8, 13] (
          <xref ref-type="bibr" rid="ref4">4</xref>
          )
Wikipedia: Used to expand definitions as discussed in section 2.2.
        </p>
        <p>Data from each source was then combined into a JSON file that totaled
approximately 300MB when compressed. All entries from each source were kept should they
contain enough information about the word, as further discussed in section 2.2. This
includes word entries that had the same (or paraphrased slightly) sense definition in
multiple sources. No merging was performed to ensure that all possible senses and
contexts were represented in the dataset. However, it may also deter the model discussed
in section 2.3 from memorizing definitions due to slight definition variations.
2.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>Modifications to data</title>
        <p>Many definitions within the training dataset did not contain enough information about
the word it was describing to be helpful to readers unfamiliar with the word. An
example of this can be seen in Figure 1A above. Here the definition of the word countries is
"plural of country." While this definition is not wrong, it does not contain enough
information for someone who does not have prior knowledge of the root word "country."
We attempt to fix this issue by identifying the referenced word and appending the
referenced word's definition to the end of the original definition.</p>
        <p>The referenced word was found using regular expressions, word frequency, and
parts-of-speech tags. Regular expressions were used to find key phrases (e.g., plural of)
that regularly pointed to the referenced word (e.g., the original word with grammatical
suffixes removed). We additionally found that many definitions contained less than four
words and heavily relied upon a key noun/adjective to convey the word's meaning. This
was appropriate from our perspective when the referenced word was frequent and
wellknown to most English speakers. However, there were also cases where the referenced
word, such as displayed in Figure 1B, was infrequent and perhaps not well-known
enough to strengthen the reader's mental representation of the original word. In cases
such as these, we also attempted to expand the definition to something more useful.</p>
        <p>To expand the definitions, we identify the referenced word and search the training
dataset for all entries of that reference word. We then use SentenceTransformer
(RoBERTa-large variant) to embed each found definition into sentence vectors, along with
the base term's contexts [16, 18]. We then compute cosine similarity to compare each
definition entry for the referenced word with the base term's contexts. The definition
with the highest cosine similarity score was then appended to the end of the definition,
as displayed in Figure 1. If the referenced word was not within the dataset, we query
Wikipedia using wptools API [24]. If the word is found, we append the first sentence
to the end of the definition. If not, the word and definition were removed from the
dataset.</p>
        <p>After removing terms with low information, the total dataset contained 254k unique
terms, 512k definitions, and 2.66M contexts.</p>
      </sec>
      <sec id="sec-2-3">
        <title>Model and Training</title>
        <p>Before training, we begin by making the data more machine-readable by placing each
word, definition, and context from the dataset into the format shown in Figure 2 for
each context entry. This format contains two special tokens: &lt;CONTEXT&gt; and
&lt;DEFINITION&gt;. These tokens were used to make it easy for the model to determine
where the context ended and where definition generation should begin.</p>
        <p>We then trained a fine-tuned version of GPT-2 called WikiMorph on the dataset [9,
22]. WikiMorph uses the large variant of GPT-2 made available by Hugging Face to
break down words into morphological compounds and definitions associated with each
compound [20]. It was chosen for this work because it contains some ability to generate
definitions for words and sub-words. Therefore, likely giving the model a good
initialization point. We then fine-tune this model on the dataset for 1 epoch. We stopped at
only a single epoch because there were no improvements in validation loss or ROUGE
scores on the validation data [17].
2.4</p>
      </sec>
      <sec id="sec-2-4">
        <title>Model Usage and Examples</title>
        <p>The model is used by placing the word and a context into the format displayed in Figure
2 and omitting everything following the special &lt;DEFINITION&gt; token. We then feed
this text into the model, which uses its tokenizer to encode the text into tensors. The
model then begins to autoregressively generate a definition by referencing the given
word, the given context, and all prior output tokens until it reaches the end token
designated as "&lt;|endoftext|&gt;". We then use regular expressions to find all text between the
&lt;DEFINITION&gt; and &lt;|endoftext|&gt; token to display to the user.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Evaluation</title>
      <sec id="sec-3-1">
        <title>Evaluation Setup</title>
        <p>
          5
To evaluate the model, we began by collecting sample texts and definitions from 5
different university-level textbooks. These were from the following subjects: Anatomy
and Physiology [6], American Government [1], Astronomy [2], Chemistry [4], and
Psychology [14]. We then searched these textbooks for paragraphs containing each of the
key terms within its glossary. These paragraphs were then divided into sentences using
Spacy's sentence segmentation model [19]. We then randomly chose 50 terms from
each textbook and generated definitions using two different context types. (
          <xref ref-type="bibr" rid="ref1">1</xref>
          ) Short
context: The sentence containing the term. (
          <xref ref-type="bibr" rid="ref2">2</xref>
          ) Long Context: The sentence containing
the word and the sentence before and the sentence after if available. We then fed the
model each of these context types to create two different Qualtrics surveys.
        </p>
        <p>
          Each survey began by displaying a multiple-choice question asking participants to
choose which of the above subjects they felt most knowledgeable in. Their selection
would determine what set of generated definitions they would assess throughout the
rest of the survey. (e.g., if they chose psychology, they would evaluate the 50 generated
definitions from the psychology textbook). For each generated definition, we asked two
questions: (
          <xref ref-type="bibr" rid="ref1">1</xref>
          ) Is the generated definition accurate in the context displayed? (
          <xref ref-type="bibr" rid="ref2">2</xref>
          ) Is the
generated definition grammatically fluent? Both questions used sliders that recorded
results on a 0-100 scale, with the starting point set at 50 for each. To see an example
question from the survey, please refer to Figure 4 above.
        </p>
        <p>We additionally created a third survey containing human-generated definitions to
compare our model against. Human-generated definitions were collected from the
training dataset described in Section 2. We searched the training dataset for all definition
entries for each term used in the prior two surveys. We then selected the most
appropriate definition using SBERT sentence vectors. Each definition for the term was
embedded into a sentence vector. We then computed cosine similarity between each of
ject.
these definitions from the training dataset with the definition from the textbook. The
definition from the training dataset with the highest similarity value was displayed to
the user in the same format shown in Figure 4. If a term happened to not be within the
training dataset, it was dropped from the survey entirely. This lowered the number of
questions for the human-generated definitions from 50 to 40-45, depending on the
sub</p>
        <p>
          For each survey, we implemented 3-4 control questions. These control questions
replaced the generated definition displayed in Figure 4 with a random definition for a
different term from the subject's textbook. Participants were required to assign an
accuracy value below 50 on the slider question for over half of the control questions to be
included in the results. (i.e., to be considered reliable raters). Below 50 was the
threshold selected due to 50 being the starting position for the slider. (i.e., the participants had
to actively move the slider towards "no" and indicate inaccuracy).
The only imposed requirement was that participants were required to be from
Englishspeaking countries. This led to a total of 194 participants. Of which, 53 were excluded
from the results for two reasons: (
          <xref ref-type="bibr" rid="ref1">1</xref>
          ) They failed to pass control checks (
          <xref ref-type="bibr" rid="ref2">2</xref>
          ) They selected
Chemistry as their preferred subject. Chemistry was removed from the results due to
having too low number of participants to calculate inter-rater reliability. Final numbers
for each group, along with inter-rater reliability, are shown in Table 1 above.
Reliability was generally high, with Cronbach's alpha mostly in the .80-.95 range.
4
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Results and Discussion</title>
      <p>
        To answer our research questions, we ran mixed-effects beta regressions for accuracy
and fluency using all three survey conditions and random intercepts for definitions and
participants. Each model was initially fit with the interaction between context and topic,
but if the interaction was not significant, it was dropped, and the model refit. For
accuracy, we ran an ANOVA and found a significant main effect of context, χ2(
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) = 42.70,
p &lt; .001. We probed this main effect to answer our first research question, which is
what context type gave the model the best initialization for definition generation. Post
hoc comparison using Tukey's HSD revealed that definitions generated from shorter
contexts (M = 62.60, SE = 0.875) performed significantly better than those generated
from longer contexts ( = 53.70,  = 0.886), t(5316)=2.38,  = 0.045. We
suspect two possible reasons as to why shorter contexts performed better in terms of
accuracy: (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) The training data had far more entries with sentence contexts (56.84% of the
total contexts) relative to longer contexts with two or more sentences (1.66% of the
total contexts), indicating the model may need more examples of longer contexts to
learn what to filter out and what to pay attention to. (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) While there could be instances
where longer contexts provide additional information, all the evaluation data came from
college-level textbooks, which may contain enough information-rich keywords within
a single sentence for the model to determine the correct sense and generate an
appropriate definition.
      </p>
      <p>
        To answer our second research question to see how the model performs relative to
human-generated definitions, we conducted an additional post hoc comparison using
Tukey’s HSD, which revealed that human-generated definitions (M = 78.57, SE = 0.70)
performed significantly better than definitions generated from shorter contexts (M =
62.60, SE = 0.875), t(5316) = 4.11, p &lt; .001, as well as those generated from longer
contexts ( = 53.70,  = 0.886), t(5316)=2.38, p &lt; .001. Further examination of
the accuracy density plot seen in Figure 5 sheds some light on this result. It shows that
the model was able to perform admirably in many situations, with a median rating equal
to 75. However, it also had an abundance of error cases where ratings were below 25.
The exact reasons for these error cases require further investigation. Two possible
culprits include: (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) The model had difficulty reading the context. As demonstrated by
Figure 6 and the fact that the model performed significantly better on short contexts, it
is safe to conclude that the model is sensitive to contexts. (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) A poor representation of
the input word due to it not being in the training data or having too many conflating
definitions without enough contextual examples to properly learn all senses.
      </p>
      <p>To answer our third research question to see if the model performed better on some
topics relative to others, we did not find significant differences between textbook
subjects. Though, a non-significant trend suggests that, while human-generated definitions
did equally well across topics, the model might perform better on some topics relative
to others. In particular, the average meaning ratings for definitions generated from
shorter contexts were approximately equal to those for human-generated definitions on
the topic of Anatomy and Physiology. This trend is potentially worthy of further
analysis in subsequent studies examining a wider variety of textbook sources and topics.</p>
      <p>We performed an identical ANOVA analysis for fluency and found no significant
main effects of context, topic, or interaction. As shown in the fluency density plot in
Figure 5, the model's performance was excellent for fluency relative to
human-generated definitions. Some of this could be due to the data found within our dataset being
slightly less fluent than a typical human definition due to some entries coming from
dirty sources or the definition expansion method discussed in Section 2.2. However,
even with this considered, the model appeared to perform exceptionally well with mean
values effectively greater than or equal to 80 for both short and long-form context
surveys.
5</p>
    </sec>
    <sec id="sec-5">
      <title>Conclusion</title>
      <p>This work presents a deep-learning model capable of dynamically generating
definitions based solely on the surrounding context. We examined the model's ability to
generate definitions using two context types: short and long-form. Short-form contexts
significantly outperformed long-form contexts in human-rated accuracy but fell short of
human-generated definitions on this metric. In contrast, short-form and long-form
conditions were indistinguishable from human-generated definitions in terms of
fluency, displaying some promise for the model.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgment</title>
      <p>This material is based upon work supported by the National Science Foundation
(1918751, 1934745) and the Institute of Education Sciences (R305A190448).</p>
    </sec>
    <sec id="sec-7">
      <title>References</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Krutz</surname>
            <given-names>G</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Waskiewicz</surname>
            <given-names>S.</given-names>
          </string-name>
          : American Government.
          <source>OpenStax</source>
          (
          <year>2018</year>
          ). https://doi.org/10.4324/9781351239226.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>Andrew</given-names>
            <surname>Franknoi</surname>
          </string-name>
          , David Morrison,
          <string-name>
            <surname>S.C.W.</surname>
          </string-name>
          : Astronomy.
          <source>OpenStax</source>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Vaswani</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shazeer</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Parmar</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Uszkoreit</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jones</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gomez</surname>
            ,
            <given-names>A.N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kaiser</surname>
          </string-name>
          , Ł.,
          <string-name>
            <surname>Polosukhin</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Attention is all you need</article-title>
          .
          <source>Adv. Neural Inf. Process. Syst. 2017-Decem</source>
          ,
          <fpage>5999</fpage>
          -
          <lpage>6009</lpage>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Flowers</surname>
            <given-names>Paul</given-names>
          </string-name>
          , Theopold Klaus, Langley Richard, Robinson William:
          <year>Chemistry 2002</year>
          . (
          <year>2002</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5. MeetDeveloper: Dictionary-Anywhere, https://github.com/meetDeveloper/Dictionary-Anywhere,
          <article-title>(</article-title>
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>David</given-names>
            <surname>Shier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Jackie</given-names>
            <surname>Butler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.L.</given-names>
            :
            <surname>Hole's Human</surname>
          </string-name>
          <string-name>
            <given-names>Anatomy</given-names>
            &amp;
            <surname>Physiology. McGraw-Hill Education</surname>
          </string-name>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Chaudhri</surname>
            , V.K., Cheng,
            <given-names>B.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Overholtzer</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Roschelle</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Spaulding</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Clark</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Greaves</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gunning</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Inquire biology: A textbook that answers questions</article-title>
          .
          <source>AI Mag</source>
          .
          <volume>34</volume>
          ,
          <fpage>55</fpage>
          -
          <lpage>72</lpage>
          (
          <year>2013</year>
          ). https://doi.org/10.1609/aimag.v34i3.
          <fpage>2486</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Miller</surname>
            ,
            <given-names>G.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Beckwith</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fellbaum</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gross</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miller</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Introduction to WordNet: An On-line Lexical Database</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Alec</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jeffrey</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rewon</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>David</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dario</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ilya</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Language Models are Unsupervised Multitask Learners</article-title>
          . (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Laufer</surname>
          </string-name>
          <article-title>Batia: Lexical Thresholds for Reading Comprehension: What They Are and How They Can Be Used for Teaching Purposes</article-title>
          .
          <volume>47</volume>
          ,
          <fpage>867</fpage>
          -
          <lpage>872</lpage>
          (
          <year>2018</year>
          ). https://doi.org/10.1 002/tesq. 1
          <fpage>40</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Lexico</surname>
          </string-name>
          . (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Robyn</surname>
            <given-names>Speer</given-names>
          </string-name>
          , And,
          <string-name>
            <given-names>J.C.</given-names>
            ,
            <surname>And</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.L.</given-names>
            ,
            <surname>And</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.J.</given-names>
            ,
            <surname>Nathan</surname>
          </string-name>
          , L.: LuminosoInsight/wordfreq, https://github.com/LuminosoInsight/wordfreq, (
          <year>2018</year>
          ). https://doi.org/10.5281/zenodo.1443582.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Bird</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Loper</surname>
          </string-name>
          , E.:
          <article-title>NLTK: The Natural Language Toolkit</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Spielman</surname>
            <given-names>Rose</given-names>
          </string-name>
          , Jenkins William, Lovett Marilyn:
          <article-title>Psychology-2e - Tâm lý học</article-title>
          .
          <source>OpenStax</source>
          (
          <year>2013</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Treptow</surname>
            ,
            <given-names>M.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Burns</surname>
            ,
            <given-names>M.K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>McComas</surname>
            ,
            <given-names>J.J.</given-names>
          </string-name>
          :
          <article-title>Reading at the frustration, instructional, and independent levels: The effects on students' reading comprehension and time on task</article-title>
          .
          <source>School Psych. Rev</source>
          .
          <volume>36</volume>
          ,
          <fpage>159</fpage>
          -
          <lpage>166</lpage>
          (
          <year>2007</year>
          ). https://doi.org/10.1080/02796015.
          <year>2007</year>
          .
          <volume>12087958</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ott</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goyal</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Du</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joshi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Levy</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lewis</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zettlemoyer</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stoyanov</surname>
          </string-name>
          , V.:
          <article-title>RoBERTa: A robustly optimized BERT pretraining approach</article-title>
          . arXiv. (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Ganesan</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>ROUGE 2.0: Updated and Improved Measures for Evaluation of Summarization Tasks</article-title>
          .
          <source>arXiv. 1-8</source>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Reimers</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gurevych</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Sentence-BERT: Sentence embeddings using siamese BERTnetworks</article-title>
          .
          <source>EMNLP-IJCNLP 2019 - 2019 Conf. Empir. Methods Nat. Lang. Process. 9th Int. Jt. Conf. Nat. Lang. Process. Proc. Conf</source>
          .
          <volume>3982</volume>
          -
          <fpage>3992</fpage>
          (
          <year>2020</year>
          ). https://doi.org/10.18653/v1/d19-
          <fpage>1410</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Honnibal</surname>
          </string-name>
          , Matthew and Montani, Ines and Van Landeghem, Sofie and Boyd, A.: spaCy:
          <string-name>
            <surname>Industrial-strength Natural</surname>
          </string-name>
          Language Processing in Python, https://spacy.io/, (
          <year>2021</year>
          ). https://doi.org/10.5281/zenodo.1212303.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Wolf</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Debut</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sanh</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chaumond</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Delangue</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cistac</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rault</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Louf</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Funtowicz</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brew</surname>
          </string-name>
          , J.: Transformers:
          <article-title>State-of-the-art natural language processing</article-title>
          . arXiv. (
          <year>2019</year>
          ). https://doi.org/10.18653/v1/
          <year>2020</year>
          .emnlp-demos.
          <volume>6</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <surname>Litman</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Robinson</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abberbock</surname>
          </string-name>
          , T.:
          <article-title>TurkPrime.com: A versatile crowdsourcing data acquisition platform for the behavioral sciences</article-title>
          .
          <source>Behav. Res. Methods</source>
          .
          <volume>49</volume>
          ,
          <fpage>433</fpage>
          -
          <lpage>442</lpage>
          (
          <year>2017</year>
          ). https://doi.org/10.3758/s13428-016-0727-z.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Yarbro</surname>
            ,
            <given-names>J.T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Olney</surname>
            ,
            <given-names>A.M.:</given-names>
          </string-name>
          <article-title>WikiMorph: Learning to Decompose Words into Morphological Structures</article-title>
          .
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>23. WikiMedia: Wiktionary, https://en.wiktionary.org.</mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>24. siznax: wptools, https://github.com/siznax/wptools.</mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>