<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>INGEOTEC at IberLEF 2019 Task HaHa</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jose Ortiz-Bejar</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Eric Tellez</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mario Graff</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Daniela Moctezuma</string-name>
          <email>dmoctezuma@centrogeo.edu.mx</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sabino Miranda-Jimenez</string-name>
          <email>sabino.mirandag@infotec.mx</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>CONACyT Consejo Nacional de Ciencia y Tecnolog a</institution>
          ,
          <addr-line>Direccion de Catedras</addr-line>
          ,
          <country country="MX">Mexico</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Centro de Investigacion en Ciencias de Informacion Geoespacial A.C.</institution>
          ,
          <country country="MX">Mexico</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>INFOTEC Centro de Investigacion e Innovacion en Tecnolog as de la Informacion y Comunicacion</institution>
          ,
          <country country="MX">Mexico</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Universidad Michoacana de San Nicolas de Hidalgo</institution>
          ,
          <country country="MX">Mexico</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2019</year>
      </pub-date>
      <fpage>203</fpage>
      <lpage>211</lpage>
      <abstract>
        <p>This manuscript describes INGEOTEC's participation in the second Humor Analysis based on Human Annotation (HAHA) task on IberLEF'2019. Our approach to solve the task was based on perform an extensive comparison of several text classifiers using a 80-20 holdout cross-validation methodology. We found that our generic text categorization and regression system ( TC) had the best performance. Finally, we conducted an analysis over the training dataset illustrating some of the task's complexity.</p>
      </abstract>
      <kwd-group>
        <kwd>Humor Analysis Text Categorization Model's Performance Analysis</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Informal writing is a complex process; it is full of ambiguity, subjective, and figurative
statements. A human learns to interpret this kind of language and expressions from
its culture and its environment. While it is quite natural for almost anybody to
understand informal language, the case of modeling it becomes a complicated task, it
is full of variants and dependencies in cultural traits that become humor identification
a challenging task.</p>
      <p>Therefore, in the end, the identification process can be tackled with a large and
diverse knowledge database, a robust enough model of the text, and a high performing
learning algorithm. However, automatic humor detection as a supervised learning
problem is complicated since the language traits that make something humorous is
hard to bound in a set of rules. It is even painful to achieve agreement among humans
on what is funny and what is not; therefore, a labeled dataset must be carefully curated.</p>
      <p>The IberLEF-2019 forum ran a task devoted to Humor Analysis based on Human
Annotation (HAHA). Here, a set of human-labeled messages from Twitter are
provided to train and test algorithms for humor identification (classification) or ranking
(regression). More detailed, each text is labeled as humorous or not humorous; a score
of the humor-intensity is also given to define a ranking problem.</p>
      <p>
        This manuscript describes the participation of INGEOTEC team in HAHA
challenge. In addition to describing our methodology and internal comparisons, we
performed an analysis of the training set to explain our performance. In particular, we
use our TC [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. The HAHA challenge is described in Section 2. Section 3 describes
our general methodology to solve the task. Section 4 is devoted to describing our
systems while experimental methodology and results are discussed in Section . Finally,
Section 7 concludes our contribution.
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>Task Description</title>
      <p>
        Humor Analysis based on Human Annotation (HAHA) [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] asks for systems that
classify tweets, in the Spanish language, as humorous or not. Also, it asks for systems
that determine how funny tweets are. Those two tasks are described by HAHA
organizers as follows:
Humor detection determining if a tweet is a joke or not (intended humor by the
author or not). The results of this task will be measured using F-measure for the
humorous category and accuracy. F-measure is the primary measure for this task.
Funniness score prediction predicting a funniness score value (average stars) for a
tweet in a 5-star ranking, supposing it is a joke. The results of this task will be
measured using the root-mean-squared error (RMSE).
      </p>
      <p>
        The first task can be solved as a classification problem, while the second one
can be tackled as a regression problem. The training set is a corpus of 24000
crowdannotated tweets, as described in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Multiple annotators evaluated each tweet, and
each annotation consists of the class (humorous or not) and the intensity (number
of stars 0-5). The final label is determined using a voting scheme. Table 2 shows an
example of the content of the provided dataset.
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>Our humor detection approach</title>
      <p>
        We select to use the modeling procedure for humor analysis described in our previous
work [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Briefly, the idea is to create a single model for both classification and
regression tasks, and it is based on computing a sparse or a dense vector space model,
and then try different learning methods that support the data model. The sparse
vector space is created through T C using an optimized text model, based on the
hyper parameter selection of the tool. For the dense modeling, we use diverse word
embedding models and then summarize word's vectors into a single dense vector.
      </p>
      <p>
        Figure 1 illustrates our generic supervised model for humor classification and
regression. For our sparse vector models we start the process with the training set
T , a set of short text messages; the text is preprocessed and tokenized using multiple
schemes like word n-grams, character q-grams, and skip-grams. This bag of tokens
is vectorized through a weighting scheme. This procedure generates the vector space
X, which can be used by a classifier or a regressor, Y depict the output associated
to the training set. The model's quality depends on the entire pipeline. The entire
process is documented in [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
      </p>
      <p>T
Y</p>
      <p>VSM</p>
      <p>X</p>
      <p>Classifier</p>
      <p>or
Regressor</p>
      <p>A similar procedure is made for dense models, that is, the text is preprocessed.
We use both pretrained word embeddings and computed word embeddings to model
our text; X is compute using the vector sum and normalization word-vectors that
compose each text.
4</p>
    </sec>
    <sec id="sec-4">
      <title>Systems description</title>
      <p>Our best solutions at both tasks were obtained using TC system. However, we
explored the use of fastText and flair along with multiple combinations of word
embeddings which range from simple character to the state-of-the-art BERT. In the
following sections, we describe several approaches in more detail as well as some
findings to hypothesize why our attempts did not show any improvement concerning
our TC baseline.</p>
      <p>
        TC [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] is a minimalist tool that generates text classifiers that maximize a
performance measurement. It manages the entire pipeline of a text classifier, as specified in
the previous section. Under the hood, TC uses a linear Support Vector Machine as the
classifier. The core idea behind TC is to define a parameter space describing a massive
number of text-classifiers. The search in this space for a competitive text classifier using
a set of heuristics to perform the search based on random search and hill climbing.
      </p>
      <p>
        We also probe EvoMSA [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], our multilingual sentiment analysis system based on
genetic programming. The core idea is the use of several and diverse models to solve
the task using a stacking scheme guided by genetic programming. This approach is
particularly robust with unbalanced classes. Besides, we tested our baseline algorithm
for multilingual sentiment analysis (B4MSA) [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]; this system is a sentiment classifier
for informal text such as Twitter messages. The design is similar to TC, but the
internal problem is solved differently, along with the use of specific features for
sentiment analysis and some language-dependent capabilities.
      </p>
      <p>
        Additionally, we also test third-party tools like FastText [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], which is a library for
text classification and word vector representation. It transforms the text into
continuous vectors that can later be used on any language related task. FastText represents
sentences with a weighted bag of words, and each word is represented as a bag of
character n-gram to create text vectors. This representation is based on the skip-gram
model [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], which take into account subword information and sharing information
across classes through a hidden representation. Also, it employs a hierarchical softmax
classifier that takes advantage of the unbalanced distribution of the classes to speed
up computation. In addition to the default configuration, we optimized many of the
parameters of FastText along with different preprocessing functions. We used random
search over a configuration space for this purpose. Finally, we also test the multilingual
library Flair [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], which implements state-of-the-art NLP models, such as named entity
recognition, part-of-speech tagging, sense disambiguation, and classification. Flair
allows to use and combine different word and document embeddings, among which
stand out flair embeddings, BERT [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] embeddings, and ELMo (see [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]) embeddings.
5
      </p>
    </sec>
    <sec id="sec-5">
      <title>Experiments and results</title>
      <p>
        We tested multiple approaches by using the tools described above. The experimental
setup consisted of using the set T of 24000 tweets human annotated by the task
organizers. Firstly, T was split in training (Tt) and validation (Tv) sets following a
80-20 proportion. Table 2 describe the validation and training sets.
and EvomSA enriched by different lexicon and decision functions learned from other
sentiment analysis tasks. Table 3 summarizes the results of our experiments.
To understand why there is no improvement over TC, we performed an analysis of
the training and validation set vocabularies to determine the similarities/differences
between them. For this propose we need to detail into the weighting scheme that
our TC uses to tackle the problem, the entropy+b term-weighting defined in [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ],
it is based on representing each term by terms entropy computed from the empirical
distribution of the available classes, using a smoothing parameter b, in this case, b = 3.
More precisely, defined as follows:
entropyb(w) = logjCj
      </p>
      <p>Xpc(w;b)log
c2C</p>
      <p>1
pc(w;b)
where C is the set of classes, and pc(w;b) is the probability of term w occurs in
class c parametrized with b. More detailed,
pc(w;b) =</p>
      <p>freqc(w)+b
b jCj+Pc2Cfreqc(w)</p>
      <p>Here freqc denotes the frequency of the term in the class c. Using this approach,
it is possible to find the set of most discriminant terms among classes; we can define
thresholds for any vocabulary size if we normalize entropyb by the logarithm of the
size's vocabulary such that terms are weighted with values between 0 to 1.</p>
      <p>Table 4 shows the sizes of the vocabularies after removing those terms with less
than 0.15 of normalized entropyb. The size of the training vocabulary is close to 80
thousand items while the vocabulary of the validation set has close 40 thousand entries.
Please recall that we used an 80-20 partition, but it is also under a combination of
tokenizers as determined by TC. To explain the findings, we refer to the terms set
(1)
(2)
for training set as model Mt and the terms for validation as Mv. The union and
intersection sizes of training and validation sets are also listed in the table.</p>
      <p>The intersection and union sizes of Table 4 indicate the number of non-shared
terms between the training and validation sets, due to it may be supposed that
semantic models will be achieve good performance, however from our experiments
using multiple state-of-the-art word embeddings do not outperformed TC model. To
gain more understand we produce Figure 2 by sorting all terms according its entropy
at the training set, it is set a value of 0 for all terms which are not in any of Mt or Mv.
the training set than in the validation set. The negative zone collects those terms with
a higher discriminant power in the training set. The zone around zero gathers terms
that have a similar entropy score at both training and validation sets. Therefore, those
terms with entropyb = 1 are highly discriminant in the training set but are not part of
the vocabulary of the validation set; conversely, terms with -1 weight are those that are
not part of the training set but exhibits unitary entropy values in the validation set.</p>
      <p>
        The previous discussion suggests that semantic models may work better; however,
pure word embeddings models achieved lower performance as described in our
experimental results. Trying to improve, we also tried kernel methods, specialized in working
with non-linear problems; in particular, we use the technique described in [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. This
approach separates the training set while does not show any significant improvement
over validation dataset scores; this behavior is an obvious symptom of overfitting.
(a) Raw vectors of the training set.
      </p>
      <p>(b) Kernelized projection of the training set.
(c) Raw vectors of the validation set.</p>
      <p>
        (d) Kernelized projection of the validation
set.
The regression sub-task was tackled using the classification model of TC but
replacing the Linear SVM classifier by the linear SVM regressor (SVR) available at
scikit-learn [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. This regressor was used to evaluate the submitted test dataset.
6
      </p>
    </sec>
    <sec id="sec-6">
      <title>Task Results</title>
      <p>Our best result using TC was ranked fifth out of 19 contestants in both, classification
and regression tasks. Table 5 shows scores for TC, the organizer's baseline and the
ones for the winner approach. For the classification task, measure F1 was used to decide
the winner, while Root Mean Squared Error (RMSE) was used for the regression task.
This paper describes the participation of the INGEOTEC team in the HAHA'19
challenge. Our final approach uses our TC system to perform both classification
and regression tasks. For this edition, we test several approaches based on different
semantic models, and our stacking solution EvoMSA; however, any of them was able
to beat TC. We performed a qualitative analysis to find possible reasons for this
situation. Possible reasons could be that our train and validation partitions contained
very different vocabulary, semantics, and high bias, as was experimentally shown.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgements</title>
      <p>The authors would like to thank to the Consorcio en Inteligencia Artificial for partially
funding this work through Project FORDECyT 296737.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Akbik</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blythe</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vollgraf</surname>
          </string-name>
          , R.:
          <article-title>Contextual string embeddings for sequence labeling</article-title>
          .
          <source>In: COLING</source>
          <year>2018</year>
          , 27th International Conference on Computational Linguistics. pp.
          <volume>1638</volume>
          {
          <issue>1649</issue>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Castro</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chiruzzo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosa</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garat</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moncecchi</surname>
          </string-name>
          , G.:
          <article-title>A crowd-annotated spanish corpus for humor analysis</article-title>
          .
          <source>In: Proceedings of SocialNLP</source>
          <year>2018</year>
          ,
          <source>The 6th International Workshop on Natural Language Processing for Social Media</source>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Chiruzzo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Castro</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Etcheverry</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garat</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Prada</surname>
            ,
            <given-names>J.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosa</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          : Overview of HAHA at IberLEF 2019:
          <article-title>Humor Analysis based on Human Annotation</article-title>
          .
          <source>In: Proceedings of the Iberian Languages Evaluation Forum (IberLEF</source>
          <year>2019</year>
          ).
          <source>CEUR Workshop Proceedings</source>
          , CEUR-WS, Bilbao,
          <source>Spain (9</source>
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Devlin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>M.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Toutanova</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Bert: Pre-training of deep bidirectional transformers for language understanding</article-title>
          . arXiv preprint arXiv:
          <year>1810</year>
          .
          <volume>04805</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Graff</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miranda-Jimenez</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tellez</surname>
            ,
            <given-names>E.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moctezuma</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Evomsa: A multilingual evolutionary approach for sentiment analysis</article-title>
          .
          <source>arXiv preprint arXiv:1812</source>
          .
          <volume>02307</volume>
          (
          <year>2018</year>
          ), https://github.com/INGEOTEC/EvoMSA
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Joulin</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grave</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bojanowski</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Bag of tricks for efficient text classification</article-title>
          .
          <source>In: Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume</source>
          <volume>2</volume>
          ,
          <string-name>
            <given-names>Short</given-names>
            <surname>Papers</surname>
          </string-name>
          . pp.
          <volume>427</volume>
          {
          <fpage>431</fpage>
          . Association for Computational Linguistics (
          <year>April 2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Mikolov</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Corrado</surname>
            ,
            <given-names>G.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dean</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          :
          <article-title>Distributed representations of words and phrases and their compositionality</article-title>
          .
          <source>In: Advances in neural information processing systems</source>
          . pp.
          <volume>3111</volume>
          {
          <issue>3119</issue>
          (
          <year>2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Ortiz-Bejar</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Salgado</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Graff</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moctezuma</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miranda-Jimenez</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tellez</surname>
            ,
            <given-names>E.S.:</given-names>
          </string-name>
          <article-title>Ingeotec at ibereval 2018 task haha: tc and evomsa to detect and score humor in texts</article-title>
          .
          <source>In: Proceedings of the Third Workshop on Evaluation of Human Language Technologies for Iberian Languages (IberEval</source>
          <year>2018</year>
          )
          <article-title>co-located with 34th Conference of the Spanish Society for Natural Language Processing (SEPLN</article-title>
          <year>2018</year>
          )
          <article-title>(</article-title>
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Ortiz-Bejar</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tellez</surname>
            ,
            <given-names>E.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Graff</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miranda-Jimenez</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ortiz-Bejar</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moctezuma</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sanchez</surname>
            ,
            <given-names>C.N.:</given-names>
          </string-name>
          <article-title>I3go+ at ricatim 2017: A semi-supervised approach to determine the relevance between images and text-annotations</article-title>
          .
          <source>In: 2017 IEEE International Autumn Meeting on Power, Electronics and Computing (ROPEC)</source>
          . pp.
          <volume>1</volume>
          {
          <issue>6</issue>
          .
          <string-name>
            <surname>IEEE</surname>
          </string-name>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Pedregosa</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Varoquaux</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gramfort</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Michel</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Thirion</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grisel</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blondel</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Prettenhofer</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weiss</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dubourg</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vanderplas</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Passos</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cournapeau</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brucher</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Perrot</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Duchesnay</surname>
          </string-name>
          , E.:
          <article-title>Scikit-learn: Machine learning in Python</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          <volume>12</volume>
          ,
          <volume>2825</volume>
          {
          <fpage>2830</fpage>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Peters</surname>
            ,
            <given-names>M.E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Neumann</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Iyyer</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gardner</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Clark</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zettlemoyer</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Deep contextualized word representations</article-title>
          .
          <source>arXiv preprint arXiv:1802</source>
          .
          <volume>05365</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Tellez</surname>
            ,
            <given-names>E.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miranda-Jimenez</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Graff</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moctezuma</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Suarez</surname>
            ,
            <given-names>R.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Siordia</surname>
            ,
            <given-names>O.S.:</given-names>
          </string-name>
          <article-title>A simple approach to multilingual polarity classification in Twitter</article-title>
          .
          <source>Pattern Recognition Letters</source>
          <volume>94</volume>
          ,
          <issue>68</issue>
          {
          <fpage>74</fpage>
          (
          <year>2017</year>
          ). https://doi.org/10.1016/j.patrec.
          <year>2017</year>
          .
          <volume>05</volume>
          .024
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Tellez</surname>
            ,
            <given-names>E.S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moctezuma</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miranda-Jimenez</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Graff</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>An automated text categorization framework based on hyperparameter optimization</article-title>
          .
          <source>Knowledge-Based Systems 149</source>
          ,
          <fpage>110</fpage>
          {
          <fpage>123</fpage>
          (
          <year>2018</year>
          ), https://github.com/INGEOTEC/microtc
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Tellez</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miranda-Jimenez</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Graff</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moctezuma</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          :
          <article-title>Gender and language-variety identification with MicroTC: Notebook for PAN at CLEF 2017</article-title>
          .
          <source>In: CEUR Workshop Proceedings</source>
          . vol.
          <year>1866</year>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Tellez</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Moctezuma</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miranda-Jimenez</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Graff</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>An automated text categorization framework based on hyperparameter optimization</article-title>
          .
          <source>Knowledge-Based Systems</source>
          (
          <year>2018</year>
          ). https://doi.org/10.1016/j.knosys.
          <year>2018</year>
          .
          <volume>03</volume>
          .003
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>