<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Applying Sentiment Analysis on Spanish Tweets Using BETO*</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ariadna de Arriba</string-name>
          <email>ariadna.de.arriba@upc.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marc Oriol</string-name>
          <email>marc.oriol@upc.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xavier Franch</string-name>
          <email>xavier.franch@upc.edu</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Universitat Politècnica de Catalunya</institution>
          ,
          <addr-line>Barcelona</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Emotion analysis of messages using machine learning techniques is a difficult and cumbersome task requiring a major effort to obtain reliable results. This challenge is even more pronounced when the target language is not English, but Spanish. To overcome this challenge, this paper describes how UPC Team applied sentiment analysis on social media messages (in particular, on Twitter) written in Spanish and, related to events that took place in April 2019 from different domains. To this aim, we present a machine learning model based on BERT and describe the results obtained to reach an accuracy of 65% approx. and the 12th position in the ranking, for this second edition of the contest for emotion detection of Spanish tweets EmoEva-lEs@IberLEF2021.</p>
      </abstract>
      <kwd-group>
        <kwd>Sentiment Analysis</kwd>
        <kwd>Machine Learning</kwd>
        <kwd>Social Media</kwd>
        <kwd>Natural Language Processing</kwd>
        <kwd>Twitter</kwd>
        <kwd>Tweets</kwd>
        <kwd>BERT</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Sentiment analysis in Spanish is a challenging task that has not been as much addressed
as in the English context. Even though the Spanish language is spoken by more than
500 million speakers (being one of the most spoken languages in the world, just behind
English, Chinese and Hindi), sentiment analysis in Spanish remains, comparatively, not
sufficiently explored.</p>
      <p>Sentiment analysis requires much time and effort to succeed in developing a good
enough machine learning model. It embraces several critical steps, from data
preprocessing (e.g., lemmatization, stemming, tokenization) to model customization (e.g.,
fine-tuning hyperparameters). But it is even more difficult when, instead of classifying
the sentiment through a polarity spectrum (from positive to negative), aims at
classifying the emotions of the messages (e.g., joy, sadness, fear, surprise).</p>
      <p>
        To overcome this issue, the Iberian Languages Evaluation Forum (IberLEF) and the
Sociedad Española para el Procesamiento del Lenguaje Natural (SEPLN) have
organized since 2012 a series of competitions and workshops to attract the attention of
researchers to develop sentiment analysis tools and techniques for the Spanish
language [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. They have always organized sentiment analysis within the polarity spectrum
analysis until past year when they introduced emotion detection. This year, they decided
to repeat the competition with emotion analysis which they named EmoEvalEs
[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>
        In this paper, we describe the system we have developed to deal with emotion
classification in Spanish tweets applying Natural Language Processing (NLP) techniques
and developing a machine learning model based on BETO [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], a Spanish version of
BERT [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>In the following section, we explain the task presented and the dataset provided.
Then, in section 3, we describe the system developed and the steps executed to
overcome the task. In the results section, we expose the metrics obtained relative to other
participants, including an analysis of our results. Finally, we close the paper with
conclusions and references.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Task description</title>
      <p>The task presented on this edition of IberLEF 2021 consists in classifying tweets into
the emotion expressed in that text.</p>
      <p>
        The dataset [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] was provided by the EmoEvalEs organizer and it is composed of
tweets written in Spanish and based on several events that took place in April 2019
related to different domains: entertainment, catastrophe, political, global
commemoration, and global strike. The corpus consists of 8223 tweets distributed in three subsets:
dev (844 tweets), train (5723 tweets) and test (1656 tweets). Dev and train corpus are
labelled with seven distinct emotions (see Table 1) and they have been used to develop
and train the machine learning model, respectively. The test subset has been used to test
the model, and its emotion distribution has been made public days after the competition
ended.
surprise (also includes distraction and amazement)
(dev)
85
16
9
181
104
eomthoetriso:nth’e emotion expressed in a tweet as ‘neutral or no 414 2800 814
      </p>
      <p>The metrics used to evaluate the performance of the task are its accuracy and the
macro averages of its precision, recall and F1 score.
3</p>
    </sec>
    <sec id="sec-3">
      <title>System description</title>
      <p>We propose a system for the classification of emotions of Tweets written in Spanish
based on a BERT variant model named BETO.</p>
      <p>To build the machine learning model, we applied the following process pipeline (see
Fig.1):
• Pre-processing tweets: In this stage we applied NLP techniques to clean and
normalize the messages in the tweets.
• Training machine learning model: Using the pre-processed tweets, we trained a
model based on BETO.
• Fine-tuning model: To improve the accuracy of our machine learning model, we
iteratively fine-tuned several hyperparameters of the BETO model and retrained the
model to obtain better results.
tweets</p>
      <p>Pre-processing
tweets</p>
      <sec id="sec-3-1">
        <title>TTrraaiinniinngg</title>
        <p>mmooddeell</p>
      </sec>
      <sec id="sec-3-2">
        <title>FFininee--ttuunniningg</title>
        <p>mmooddeell</p>
        <p>
          Final
model
Pre-processing is a critical step of all NLP systems [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]. We applied some general NLP
pre-processing techniques and specific methods for emotion processing to obtain a
“cleaner” text. Some examples of tweets with their preprocessing results are shown in
Table 2. The methods and techniques applied are:
• Remove URLs: Many tweets contain URLs to websites that provide more
information but that are not relevant to the sentiment analysis process. We removed these
URLs as they only make the text noisier.
• Remove hashtags: We removed all hashtags in the text.
• Remove numbers: Numbers are not providing useful information for emotion
analysis.
• Replace emojis and emoticons: Emojis and emoticons are remarkably significant
in sentiment analysis. A text can be apparently neutral but can express any emotion
by adding emojis. To facilitate the machine learning process, we replaced the emojis
and emoticons with a text that represents the emotion they evoked.
• Replace abbreviations: Social media users express themselves in a colloquial
mode. These users tend to use abbreviated words and expressions that are not usually
present in dictionaries for NLP but are known to everybody. To this aim, we built a
list of fifty common abbreviations and we replaced them with their correct form.
Some of them are: q (que), bn (bien) o mñn (mañana), tqm (te quiero mucho), pti
(para tu información).
• Replace laughs: Laughs may be ambiguous but they are mostly used to express the
‘happy’ emotion. Nevertheless, users tend to write laughs in multiple forms. So we
replaced all words that start with 'ja', 'je', 'ji', 'jo' with 'jajaja' (after checking that the
term does not exist in the dictionary to avoid replacing existing words), to obtain a
general form for laughing.
• Remove punctuation marks: We removed punctuation marks such as commas,
question marks, or quotation marks because they do not provide any emotion in the
text by themselves. In several cases, some punctuation marks (e.g., exclamation
marks) can emphasize the emotion that the text expresses. As there is no generalized
rule for making the machine detect these cases in a clear way, we decided to remove
them prior to producing more confusion to the machine.
• Remove repeated characters: In social media, most people are not following
spelling rules and it is common to repeat some characters especially at the end of a
word (e.g. holaa or graciaas!). We removed these additional characters.
• Lemmatization: Common technique in NLP. It consists in transforming all verb
conjugations into its infinitive form. For example, we replaced vivían by vivir.
• Remove stopwords: Stopwords are words that are not providing any useful
information (e.g., y, aunque, con). Removing them is very common in NLP. In Python
there are many libraries to obtain a list of stopwords in different languages. In this
case, we used the nltk library [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ].
• Remove blank spaces: The last stage is removing extra white spaces that tweets
could have or the ones created when we replaced or deleted words in previous steps.
Some of these rules have not been tested in this dataset (e.g., repeated characters
removal or laughs removal) due to the messages are not containing any of them. Even so,
we have included it as it is very common in social media users.
La mejor manera de entender mi idioma #DiaDelLibro
https://t.co/W9wwoyW8qk
Espero que todos ya estén listos para gritar los goles
del Barça!! porque hay que creer y confiar que mañana
ellos van a ganar ♥⚽⚽
#ChampionsLeague #Barça
        </p>
        <p>Preprocessed tweet
devastador tristeza no parecer ir
quedar mucho
mejor manera entender idioma
esperar listo gritar gol haber creer
confiar mañana ir ganar felicidad
felicidad
3.2</p>
        <sec id="sec-3-2-1">
          <title>Training</title>
          <p>
            We trained the machine learning model using a Spanish variant of the BERT model
named BETO. BERT is a machine learning technique pre-trained with the Toronto
Book Corpus and Wikipedia [
            <xref ref-type="bibr" rid="ref4">4</xref>
            ] and developed by Google [
            <xref ref-type="bibr" rid="ref8">8</xref>
            ]. BERT is a
transformerbased deep learning model able to deal with multiple NLP problems. It uses attention
masks to encode each word during the training stage and predict up to 15% of masked
words using NSP (Next Sentence Prediction) and to understand the context of a
sentence [
            <xref ref-type="bibr" rid="ref9">9</xref>
            ].
          </p>
          <p>
            BETO is a variant of BERT, which has been pre-trained exclusively on Spanish data,
with a dataset of similar size as BERT. We have chosen BETO over BERT, as BETO
has been able to outperform other BERT-based models in several NLP-based activities
using Spanish as language [
            <xref ref-type="bibr" rid="ref10">10</xref>
            ].
          </p>
          <p>During the training stage, some configuration parameters are fixed meanwhile others
are adjusted as detailed in the next section. The parameters set before starting training
the model are the batch size and the maximum sequence length. The batch size is the
number of samples used in one iteration in the training stage and was set to 64 . The
maximum sequence length was set to 256 as tweets are short texts up to 240 characters
maximum.
3.3</p>
        </sec>
        <sec id="sec-3-2-2">
          <title>Fine-tuning</title>
          <p>In the fine-tuning stage, we adjusted several hyperparameters (see Table 3) to improve
our trained model and obtain better results.</p>
          <p>
            The hyperparameters we adjusted are:
• Learning rate: Hyperparameter that controls how much the model changes in
response to the estimated error each time the model weights are updated. Choosing the
optimal learning rate is a difficult task, as a small learning rate may result in a slow
training process and a value that is too large can cause the model to diverge instead
of to converge to the solution (see Fig. 2) [
            <xref ref-type="bibr" rid="ref11">11</xref>
            ].
• Epsilon: This hyperparameter is a very small number to prevent any division by zero
in the implementation [
            <xref ref-type="bibr" rid="ref13">13</xref>
            ].
• Number of epochs: Hyperparameter that indicates the number of times that the
model visits the entire training dataset. We adjust it to control the weight decay as it
uses the following formula:
weight decay = learning rate / number of epochs
(1)
In this regard, we should be careful choosing the value for epochs as weight decay
is used to prevent overfitting and to keep a weight small to avoid exploding
gradients.
The results extracted from the evaluation phase are shown in Table 4. We submitted
three different versions trained with different hyperparameters tuned. The first
submission was trained with 15 epochs, a learning rate of 10-5, and epsilon of 10-4. In the
second one, we decided to reduce the epsilon to 10-5 but keep the learning rate and the
number of epochs to check if model performance improves changing only the epsilon
value. Finally, in the third submission, epsilon and learning rate were set to 10-4
meanwhile the number of epochs was reduced to 10.
#
1
2
3
4
          </p>
          <p>Learning rate
1x10-5
1x10-5
1x10-4</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Results</title>
      <p>Epsilon
1x10-4
1x10-5
1x10-4</p>
      <p>Number of epochs
15
15
10
#
1
2
3</p>
      <p>Accuracy
As the best performance results correspond to the second submission, that is the one we
submitted to the official leaderboard. In relation with other participants, we obtained an
accuracy of 65% approximately which placed us in the 12th position in the ranking.
Despite this position, the accuracy obtained is not so far from the winner which has got
around a 73% of this metric.</p>
      <p>As we can observe, the F1 score is very close to accuracy, which indicates that the
dataset is sufficiently balanced. Metrics results are very similar in all submissions
which may indicate that model cannot be improved so much only by fine-tuning the
hyperparameters chosen with the data provided.</p>
      <p>The main errors we have detected in tweets classification in the validation stage are
mainly due to pre-processing issues. For instance, we identified that tweets that may be
predicted as ‘others’ are always classified as ‘sadness’ or ‘joy’, probably by the
influence of replacing emojis in the pre-processing phase. Another issue is that the system
is less accurate for detecting emotions that have a small number of samples, as it is the
case for ‘fear’, ‘disgust’ and ‘surprise’.</p>
      <p>For future work, we could obtain more data from Twitter to increase the performance
of the model by training it with a bigger size corpus and a little bit more balanced. In
this case, we only trained the model with the train dataset but we could have trained it
with train and dev dataset to have more data for the machine. Another improvement for
the future could be trying other values and combinations for the hyperparameters that
cannot be tested for this competition or adjusting other parameters as the batch size.
5</p>
    </sec>
    <sec id="sec-5">
      <title>Conclusions</title>
      <p>Developing a machine learning model to classify a text into emotions is a challenging
task. The way human beings express themselves is very ambiguous and detecting which
emotions they want to express can be extremely difficult even for them, because many
times they do not even know how they feel. For this reason, it is important to have good
quality data with labelled emotions well remarked and above all be patient as finding
an optimal model is a slow and hard task.</p>
      <p>In this paper, we have presented a BETO-based machine learning model to classify
into emotions Spanish tweets. The results of the evaluation show that the model is able
to identify the correct emotion on approximately 2 out of 3 occasions (accuracy=0.65).
As future work, we plan to improve the overall system by enhancing the preprocessing
phase (e.g., taking into account capital letters that could emphasize the text emotion or
keeping some hashtags that could be valuable for our aim), fine-tuning further the
hyperparameters of the machine learning model and make an in-depth quantitative error
analysis to improve our results in emotion classification.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Montes</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gonzalo</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aragón</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Agerri</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Álvarez-Carmona</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Á</surname>
          </string-name>
          ., …
          <string-name>
            <surname>Taulé</surname>
          </string-name>
          , M. (Eds.). (
          <year>2021</year>
          ).
          <source>Proceedings of the Iberian Languages Evaluation Forum (IberLEF</source>
          <year>2021</year>
          ).
          <source>CEUR Workshop Proceedings.</source>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Plaza-del-Arco</surname>
            ,
            <given-names>F. M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jiménez-Zafra</surname>
            ,
            <given-names>S. M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Montejo-Ráez</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Molina-González</surname>
            ,
            <given-names>M. D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ureña-López</surname>
            ,
            <given-names>L. A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Martín-Valdivia</surname>
            ,
            <given-names>M. T.</given-names>
          </string-name>
          (
          <year>2021</year>
          ).
          <article-title>Overview of the EmoEvalEs task on emotion detection for Spanish at IberLEF 2021</article-title>
          .
          <source>Procesamiento del Lenguaje Natural</source>
          ,
          <volume>67</volume>
          (
          <issue>0</issue>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Cañete</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chaperon</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fuentes</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ho</surname>
            ,
            <given-names>J.-H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Pérez</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          (
          <year>2020</year>
          ).
          <article-title>Spanish PreTrained BERT Model and Evaluation Data</article-title>
          .
          <source>In PML4DC at ICLR</source>
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Devlin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chang</surname>
          </string-name>
          , M.-W.,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Toutanova</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          (
          <year>2019</year>
          ).
          <article-title>BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding</article-title>
          . arXiv:
          <year>1810</year>
          .04805 [cs]. Retrieved from http://arxiv.org/abs/
          <year>1810</year>
          .04805
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Plaza-del-Arco</surname>
            ,
            <given-names>F. M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Strapparava</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ureña-López</surname>
            ,
            <given-names>L. A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Martín-Valdivia</surname>
            ,
            <given-names>M. T.</given-names>
          </string-name>
          (
          <year>2020</year>
          ).
          <article-title>EmoEvent: A Multilingual Emotion Corpus based on different Events</article-title>
          .
          <source>In Proceedings of the 12th Language Resources and Evaluation Conference</source>
          (pp.
          <fpage>1492</fpage>
          -
          <lpage>1498</lpage>
          ). Marseille, France: European Language Resources Association. Retrieved from https://www.aclweb.org/anthology/2020.lrec-
          <volume>1</volume>
          .
          <fpage>186</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Uysal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Günal</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          (
          <year>2014</year>
          ).
          <article-title>The impact of preprocessing on text classification</article-title>
          .
          <source>Inf</source>
          . Process. Manag.,
          <volume>50</volume>
          ,
          <fpage>104</fpage>
          -
          <lpage>112</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <source>Natural Language Toolkit - NLTK 3.6.2 documentation. (n.d.)</source>
          .
          <source>Retrieved May 28</source>
          ,
          <year>2021</year>
          , from https://www.nltk.org/
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Vaswani</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shazeer</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Parmar</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Uszkoreit</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jones</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gomez</surname>
            ,
            <given-names>A. N.</given-names>
          </string-name>
          , …
          <string-name>
            <surname>Polosukhin</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          (
          <year>2017</year>
          ).
          <article-title>Attention Is All You Need</article-title>
          . arXiv:
          <volume>1706</volume>
          .03762 [cs]. Retrieved from http://arxiv.org/abs/1706.03762
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Horev</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          (
          <year>2018</year>
          , November 17). BERT Explained:
          <article-title>State of the art language model for NLP</article-title>
          .
          <source>Medium. Retrieved February 12</source>
          ,
          <year>2021</year>
          , from https://towardsdatascience.com
          <article-title>/bert-explainedstate-of-the-art-language-model-for-nlp-f8b21a9b6270</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <article-title>Plaza-del-</article-title>
          <string-name>
            <surname>Arco</surname>
            ,
            <given-names>F. M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Molina-González</surname>
            ,
            <given-names>M. D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ureña-López</surname>
            ,
            <given-names>L. A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Martín-Valdivia</surname>
            ,
            <given-names>M. T.</given-names>
          </string-name>
          (
          <year>2021</year>
          ).
          <article-title>Comparing pre-trained language models for Spanish hate speech detection</article-title>
          .
          <source>Expert Systems with Applications</source>
          ,
          <volume>166</volume>
          , 114120. https://doi.org/10.1016/j.eswa.
          <year>2020</year>
          .114120
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Brownlee</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          (
          <year>2019</year>
          , January 22).
          <article-title>How to Configure the Learning Rate When Training Deep Learning Neural Networks</article-title>
          .
          <source>Machine Learning Mastery. Retrieved</source>
          from https://machinelearningmastery.com
          <article-title>/learning-rate-for-deep-learning-neural-networks/</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <article-title>Setting the learning rate of your neural network</article-title>
          .
          <source>(2018, March</source>
          <volume>2</volume>
          ).
          <source>Jeremy Jordan. Retrieved May 31</source>
          ,
          <year>2021</year>
          , from https://www.jeremyjordan.me/nn-learning-rate/
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Brownlee</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          (
          <year>2017</year>
          , December 19).
          <article-title>A Gentle Introduction to Transfer Learning for Deep Learning</article-title>
          .
          <source>Machine Learning Mastery. Retrieved</source>
          from https://machinelearningmastery.com
          <article-title>/transfer-learning-for-deep-learning/</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>