<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>HAHA at EmoEvalEs 2021: Sentiment Analysis in Spanish Tweets with Cross-lingual Model</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Kun Li</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>School of Information Science and Engineering, Yunnan University</institution>
          ,
          <addr-line>Yunnan</addr-line>
          ,
          <country country="CN">P.R. China</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This article describes HAHATeam participation in the IberLEF 2021 EmoEvalEs task: Emotion detection and Evaluation for Spanish. In this task, we use Masked Language Model-based data augmentation to enhance the data to increase the training data set and prevent over tting. We improve the performance of the model through a series of data processing and data augmentation technologies. We use three di erent cross-language models, BERT, XLM, and XLM-RoBERTa for comparative experiments. Very competitive results have been achieved on both the development set and the test set. The best model achieved the third place in the development set and the fth place in the test set.</p>
      </abstract>
      <kwd-group>
        <kwd>Sentiment Analysis</kwd>
        <kwd>Data Augmentation</kwd>
        <kwd>Machine Learning</kwd>
        <kwd>Cross-lingual Model</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Sentiment analysis (SA) is the process of analyzing, processing and categorizing
subjective text. According to the emotional tendency of the text, the text can be
classi ed. In addition to the traditionally recognized positive and negative
emotions, there should also be anger, disgust, fear, joy, sadness, surprise and other
emotion types. This article describes the emotion classi cation task (EmoEvalEs
[
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]) that we participated in in IberLEF 2021 [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], which requires the classi
cation of Spanish tweets into various categories based on emotions, such as anger,
disgust, fear, joy, sadness, surprise, or others. Due to the lack of voice regulation
and facial expressions, understanding the emotions expressed by users on social
media is a di cult task.
      </p>
      <p>
        The earliest work on SA largely relied on feature engineering [
        <xref ref-type="bibr" rid="ref15 ref3">15,3</xref>
        ], and
subsequent neural network-based methods [
        <xref ref-type="bibr" rid="ref14 ref16 ref17 ref8">8,17,14,16</xref>
        ] achieved higher accuracy.
Recently, Ma [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] said: \Incorporate useful common sense knowledge into deep
neural networks to further enhance the results of the model." Liu [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] optimized
the memory network and applied it to its model to better capture the language
structure.
      </p>
      <p>
        Recently, pre-trained language models, such as ELMo [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ], OpenAIGPT [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ],
and BERT [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], have shown their e ectiveness in mitigating feature engineering
e orts. In particular, BERT has achieved excellent results in Question Answering
(QA) and Natural Language Inference (NLI). However, the SA task that directly
uses the pre-trained BERT model does not improve much.
      </p>
      <p>In order to solve this task, we proposed the use of pretrained classic deep
learning models to apply to downstream task. Furthermore, we not only tested
the data on a single model, but also compared three language models. We use
the three mainstream cross-language models: bert-multiling-cased (BERT),
xlmmlm-100-1280 (XLM), xlm-roberta-base (XLM-RoBERTa). These three models
all come from the transformers package, and the hugging face provided us with
great help. The system will sort according to accuracy and the weighted-averaged
versions of Precision, Recall, and F1.</p>
      <p>The rest of the article is structured as follows. Section2 proposes the task to
be solved and introduces the corpus of the task. It analyzes the data set and
proposes some tricks. In the Section3, we introduce the methods and techniques
we used. The rst is to use some conventional operations to process the data
set, the second is to introduce the data augmentation method we use, and the
third is to introduce the language model we use. Section 4 presents the results
and discuss the performance of the system. In Section 5, we give the conclusions
and future work.
2
2.1</p>
    </sec>
    <sec id="sec-2">
      <title>Task and Corpus description</title>
      <sec id="sec-2-1">
        <title>Task description</title>
        <p>
          The organizer only proposed one task: EmoEvalEs [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]: Emotion detection and
Evaluation for Spanish. Understanding the emotions expressed by users on
social media is a hard task due to the absence of voice modulations and facial
expressions. Our shared task \Emotion detection and Evaluation for Spanish"
has been designed to encourage research in this area. The task consists of
classifying the emotion expressed in a tweet as one of the following emotion classes:
Anger, Disgust, Fear, Joy, Sadness, Surprise or Others.
2.2
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>Corpus description</title>
        <p>
          The datasets [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] is based on events related to di erent elds that occurred
in April 2019: entertainment, disasters, politics, global memorials, and global
strikes. The corpus contains approximately 8.2k tweets for this task. Each tweet
contains 5 features: id, event, tweet, o ensive and emotion. The most important
feature is tweet. At the same time, event and o ensive may also have a positive
impact on the nal result. We will verify our conjecture in the next part of
the experiment. For this task, the corpus is released as three data sets, namely
training, development and testing. Table 1 shows the distribution of training set
and development set.
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Method and technology</title>
      <sec id="sec-3-1">
        <title>Data Preprocessing</title>
        <p>
          To achieve good results, data preprocessing is essential. The rst step in our
model is to process these tweets. For how to preprocess the data, we refer to
IberLEF 2019 [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ], which gives us some ideas and and more in-depth analysis.
We found that in addition to tweets, there are two other features that may a ect
that nal result, so we process the datasets as follows:
{ Combining tweet, event, and o ensive together as a new text. The advantage
of this is to make better use of the information given by the organizer.
{ Removing URLs.
{ Contracting white-space characters, such as space, tabs or punctuation.
{ Removing stop word in Spanish.
        </p>
        <p>We ignored spelling errors in text tweets. After data processing, the sentence
length in the tweet is distributed within 60. At the same time, we also keep the
version that not remove the stop word in Spanish, so that we can judge whether
removing the stop word in Spanish is good for our model.
3.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>Data Augmentation</title>
        <p>Data augmentation technology is widely used in computer vision, but it is rarely
e ectively used in Natural Language Processing (NLP). The essential reason is
that some data augmentation methods in the image will not change the meaning
of the image itself while augmenting the data. In this article, we try to use a
data augmentation method: Masked Language Model-based data augmentation.
Masked Language Modeling (MLM) can not only predict words, but also can be
used for text data augmentation, and other methods, such as based on synonym
substitution and word embedding. Compared with the replacement of, the text
generated by data augmentation based on the Masked Language Model is more
grammatically uent, because the model considers the context information when
making predictions. We can easily use HuggingFace's transformers library. One
place to pay attention to in this method is how to determine which part of the
Mask text is to be used. Do data augmentation for each processed tweet, and
determine the number of masks according to the length of each sentence, details
as follows:
{ Fixed 1 [Mask], random position, repeatable.
{ Fixed 2 [Mask], random position, repeatable.
{ Fixed 3 [Mask], random position, repeatable.
{ 15% of the length of the sentence [Mask], the position is random, not
repeatable.
{ 20% of the length of the sentence [Mask], the position is random, not
repeatable.</p>
        <p>Next is a concrete example: `This is [Mask] cool'
`score': 0.515411913394928, `sequence': `This is [pretty] cool'
`score': 0.1166248694062233, `sequence': `This is [really] cool'
`score': 0.07387523353099823, `sequence': `This is [super] cool'
`score': 0.04272908344864845, `sequence': `This is [kinda] cool'</p>
        <p>\score" indicates the degree of similarity between the word and [Mask], and
\sequence" indicates the sequence after data augmentation. The sequence with
the highest score is selected as the sentence augmented by our data.
3.3</p>
      </sec>
      <sec id="sec-3-3">
        <title>Model</title>
        <p>For SA Task, we combine some classic deep learning models. In the deep learning
model, we mainly use the most popular cross-lingual models. These cross-lingual
models can be better used to process Spanish text. These models include two
parts: tokenizer and SequenceClassi cation. Among them, the tokenizer divides
the input into words and processes it into the input required by the model.
SequenceClassi cation is applied to speci c downstream tasks.</p>
        <p>{ tokenizer: First of all, the tokenizer was splitting strings in sub-word token
strings, converting tokens strings to ids and back, and encoding or
decoding. Adding new tokens to the vocabulary in a way that is independent of
the underlying structure (Byte Pair Encoding). Managig special tokens (like
mask, begining-of-sentence, etc). The model needs language embedding
during inference. We will use the tokenizer module to process the input before
entering the model. The input required for each cross-language model is
different, and the tokenizer module will generate the corresponding embedding
according to the speci c model. The tweet rst undergoes data
preprocessing, and then uses the tokenizer to vectorize the input. According to di erent
models, Token embeddings, Segment Embeddings, and Position Embeddings
are obtained.
{ SequenceClassi cation: The model with a sequence classi cation/regression
heads on top(a linear layer on top of the pooled output). We add a dropout
layer and a linear layer to the last layer. The input feature of the linear layer
we added is 768, the output feature is 384, and the p value of the dropout
layer is set to 0.1. The input feature of the last layer is 384, and the output
feature is 7, to satisfy our 7 classi cation task. The model does not require
language embedding at inference time. They can identify the language used
in the context and infer accordingly.</p>
        <p>
          XLM-RoBERTa [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]: The XLM-RoBERTa model was proposed in
Unsupervised Cross-lingual Representation Learning. It is a large multi-lingual language
model, trained on 2.5TB of ltered Common Crawl data. XLM-RoBERTa is a
multilingual model trained on 100 di erent languages. Unlike some XLM
multilingual models, it does not require lang tensors to understand which language
is used, and should be able to determine the correct language from the
inputids. XLM-RoBERTa was evaluated on classi cation tasks and they showed very
good performance. During ne-tuning, multi-language annotation data is used
based on the ability of the multi-language model to improve the performance
of downstream tasks. This allows our model to obtain state-of-the-art results in
cross-language benchmark tests, while surpassing the performance of the
singlelanguage BERT model in each language. The main body of the XLM-RoBERTa
model is Transformer, and the training target is the multilingual MLM objective,
which is basically the same as XLM. We sample the text from each language,
and then predict the tokens that are masked. We use model Sentence Piece with
unigram language model to perform Sentence Piece on the original text data.
        </p>
        <p>
          XLM [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ]: Although the BERT model can be pre-trained on hundreds of
languages, the information between languages is not interoperable, and there is no
shared knowledge between di erent language models. Facebook's XLM model
overcomes the problem of non-interoperability of information, puts di erent
languages together and uses new training targets for training, so that the model
can master more cross-language information. A signi cant advantage of this
cross-language model is that, for subsequent tasks after pre-training (such as
text classi cation or translation tasks), languages with relatively rare training
corpus can use the information learned on other corpora. We propose to use
xlm-mlm-100-1280 as part of our experiment.
        </p>
        <p>
          BERT: The bert-base-multilingual-cased model is based on BERT, which has
obtained state-of-the-art performance on most NLP task [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. The pre-trained
BERT models are trained on a large corpus (Wikipedia). Trained on cased text
in the top 104 languages with the largest Wikipedias. So
bert-base-multilingualcased model is better for Spanish text in Sentiment Analysis (SA). Because
bertbase-multilingual-cased model can splits tokens in its vocabulary into sub-tokens,
while will a ect the result of the classi cation task. The
bert-base-multilingualcased model also xes normalization issues in many languages, so it is
recommended in languages with non-Latin alphabets (and is often better for most
languages with Latin alphabets).
4
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Experiments</title>
      <p>We have fully experimented with the above three model. The process and
result of the experiment will be described in detail below. We rst experiment
with di erent data augmentation methods to obtain better data augmentation
methods. Then we use the best data augmentation method on the three
crosslanguage models and compare them on the development set. Finally, we use the
best model to experiment on the test set and get the nal result.
4.1</p>
      <sec id="sec-4-1">
        <title>Experimental Setup</title>
        <p>During the development set experiment, we divided the training set into a new
training set and a development set, where the division ratio was 0.9 for the new
training set, and the development set was used as the test set.</p>
        <p>We will do the same preprocessing before sending the data to each model
to ensure the comparability of the experiment, that is the ablation experiment.
Use ablation experiments to evaluate the quality of our di erent data processing
methods. And here some hyper parameter sett of the model; the max length of
a single sentence in each batch is adjusted according to the preprocessed results,
the max length does not exceed 60; and add \[CLS]" and \[SEP]" to each input;
and padding for sentences with insu cient length; We use a batch size of 32, a
base learning rate of 1e-5; the activation function is AdamW; and the number
of ne-tuning is set to 3. For every 10 batch-size training set data processed,
model.eval will be executed to process the development set data. And save the
model and print the accuracy rate.
4.2</p>
      </sec>
      <sec id="sec-4-2">
        <title>Learning Rate Warm-up</title>
        <p>Use learning rate warm-up. During warm-up, the learning rate linearly increases
from 0 to the initial learning rate of the optimizer. After the warm-up phase, a
schedule is created to linearly reduce the learning rate from the initial learning
rate in the optimizer to 0. The number of steps for the warm up phase is set to
0, that is, the learning rate is warmed up from the beginning. The total number
of training steps are set to N.</p>
        <p>N = len(train dataloader)
num epochs
(1)
4.3</p>
      </sec>
      <sec id="sec-4-3">
        <title>Data Augmentation Experiments</title>
        <p>We conduct data augmentation experiments on the XLM-RoBERTa model to
get the best data augmentation method, and use the data augmentation method
in all subsequent models. From Table 2, we can see that when using 15% of
the length of the sentence [Mask], XLM-RoBERTa is the best among the three
evaluation indicators Accuracy, Recall weighted and F1 weighted score. But
when the XLM-RoBERTa model uses 20% of the sentence length [Mask], the
precision of the model is the best. The choice of pre-training model has a great
impact on the nal performance, even though they are all based on transformers.
4.4</p>
      </sec>
      <sec id="sec-4-4">
        <title>Finally Experiments</title>
        <p>Next, we conduct ablation experiments on three cross-language models for data
augmentation. From Table 3, we can see that the four evaluation indicators</p>
        <p>Data Augmentation Accuracy Precision weighted Recall weighted F1 weighted
Yes
No
Yes
No
Yes
No
of the BERT model and the XLM model during the data augmentation
operation are less than those without the data augmentation operation. In the
XLM-RoBERTa model, data augmentation can improve the performance of the
model to a certain extent, except for Precision weighted scores. Furthermore,
whether data augmentation is performed or not, the performance of the
XLMRoBERTa model is better than the other two models. The biggest advantage of
XLM-RoBERTa is the signi cant increase in the amount of training data. And
the e ect is especially obvious in languages with fewer resources.</p>
        <p>Finally, we use the best performing XLM-RoBERTa model on the test data
set, with data augmentation and no data augmentation at the same time. From
Table 4, we can see that data augmentation will e ectively improve the
performance of the model in test data set, especially in Accuracy, Recall weighted and
F1 weighted scores. But it will also reduce the Precision weighted score of the
model.
We use three popular cross-language models BERT, XLM, and XLM-RoBERTa
for comparative experiments. It can be seen from Figure 3 that the
XLMRoBERTa model surpasses BERT and XLM in all evaluation indicators. Because
the XLM-RoBERTa model increases the number of languages and the number
of training data sets, it uses more than 2TB of pre-processed CommonCrawl
data sets. And during the ne-tuning of the XLM-RoBERTa model, based on
the ability of the multi-language model to use multi-language annotation data
to improve the performance of downstream tasks. Beyond this, although we use
data augmentation technology to only double the amount of data, the
performance improvement is still very obvious. Accuracy and Recall weighted have
increased by 0.6%, and F1 weighted has increased by 1.2%. Moreover, the data
processing method and Learning Rate Warm-up technology we use also play a
certain role in the improvement of model performance and the speed of model
inference.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusions and further work</title>
      <p>This article introduces the EmoEvalEs task: Emotion detection and Evaluation
for Spanish in the IberLEF 2021. We use three cross-language models BERT,
XLM and XLM-RoBERTa. Using the cross-language pre-training model can
handle Spanish tweets well, and the pre-training model only needs to modify the
nal output layer to be well applied to downstream tasks, which greatly reduces
the training time. We improve the performance of the model through a series of
data processing and data augmentation technologies. As can be seen from Table
5, The results and ranking of our model on the o cial test sets (Ranked by
Accuracy). Our nal model got an Accuracy of 0.692, a Precision weighted score
of 0.680, a Recall weighted score of 0.692, and an F1 weighted score of 0.664
on the o cial test set, ranking 5th, 6th, 5th, and 8th among all participating
teams.</p>
      <p>We are very satis ed with our rst participation in IberLEF 2021 and our
performance in the EmoEvalEs task, but some work still needs improvement.
The future work is to explore more advanced technology, use better models,
and achieve better results in the next competition. For example, use better data
augmentation technology to increase the amount of data set, select the best
hyperparameter adjustment technology, and use other pre-training models such
as Longformer, ELECTRA, ALBERT or even T5.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>First of all, we would like to thank the organizers for the opportunity and
organization, as well as the help of teachers and seniors. Finally, we would like
to thank the school for supporting my research and the patient work of future
reviewers.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Conneau</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Khandelwal</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goyal</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chaudhary</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wenzek</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Guzman</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grave</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ott</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zettlemoyer</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stoyanov</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Unsupervised cross-lingual representation learning at scale</article-title>
          . CoRR abs/
          <year>1911</year>
          .02116 (
          <year>2019</year>
          ), http://arxiv. org/abs/
          <year>1911</year>
          .02116
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Devlin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Toutanova</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>BERT: pre-training of deep bidirectional transformers for language understanding</article-title>
          . CoRR abs/
          <year>1810</year>
          .04805 (
          <year>2018</year>
          ), http://arxiv.org/abs/
          <year>1810</year>
          .04805
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Kiritchenko</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhu</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cherry</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mohammad</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <string-name>
            <surname>NRC-Canada-</surname>
          </string-name>
          2014:
          <article-title>Detecting aspects and sentiment in customer reviews</article-title>
          .
          <source>In: Proceedings of the 8th International Workshop on Semantic Evaluation (SemEval</source>
          <year>2014</year>
          ). pp.
          <volume>437</volume>
          {
          <fpage>442</fpage>
          . Association for Computational Linguistics, Dublin, Ireland (Aug
          <year>2014</year>
          ). https://doi.org/10.3115/v1/
          <fpage>S14</fpage>
          -2076, https://www.aclweb.org/ anthology/S14-2076
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Lample</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Conneau</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Cross-lingual language model pretraining</article-title>
          . CoRR abs/
          <year>1901</year>
          .07291 (
          <year>2019</year>
          ), http://arxiv.org/abs/
          <year>1901</year>
          .07291
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cohn</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Baldwin</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Recurrent entity networks with delayed memory update for targeted aspect-based sentiment analysis</article-title>
          .
          <source>In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>2</volume>
          (Short Papers). pp.
          <volume>278</volume>
          {
          <fpage>283</fpage>
          . Association for Computational Linguistics, New Orleans,
          <source>Louisiana (Jun</source>
          <year>2018</year>
          ). https://doi.org/10.18653/v1/
          <fpage>N18</fpage>
          -2045, https://www.aclweb.org/ anthology/N18-2045
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Ma</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Peng</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cambria</surname>
          </string-name>
          , E.:
          <article-title>Targeted aspect-based sentiment analysis via embedding commonsense knowledge into an attentive lstm (</article-title>
          <year>2018</year>
          ), https://www. aaai.org/ocs/index.php/AAAI/AAAI18/paper/view/16541
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Montes</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gonzalo</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aragon</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Agerri</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alvarez-Carmona</surname>
            ,
            <given-names>M.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Alvarez Mellado</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Carrillo-de Albornoz</surname>
          </string-name>
          , J.,
          <string-name>
            <surname>Chiruzzo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Freitas</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gomez</surname>
            <given-names>Adorno</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            ,
            <surname>Gutierrez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            ,
            <surname>Jimenez-Zafra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.M.</given-names>
            ,
            <surname>Lima</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Plaza-de Arco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.M.</given-names>
            ,
            <surname>Taule</surname>
          </string-name>
          , M. (eds.):
          <source>Proceedings of the Iberian Languages Evaluation Forum (IberLEF</source>
          <year>2021</year>
          ) (
          <year>2021</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Nguyen</surname>
            ,
            <given-names>T.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shirai</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>PhraseRNN: Phrase recursive neural network for aspect-based sentiment analysis</article-title>
          .
          <source>In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing</source>
          . pp.
          <volume>2509</volume>
          {
          <fpage>2514</fpage>
          . Association for Computational Linguistics, Lisbon, Portugal (Sep
          <year>2015</year>
          ). https://doi.org/10.18653/v1/
          <fpage>D15</fpage>
          -1298, https://www.aclweb.org/anthology/ D15-1298
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Pastorini</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pereira</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zeballos</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chiruzzo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rosa</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Etcheverry</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Retuyt-inco at tass 2019: Sentiment analysis in spanish tweets</article-title>
          . In: IberLEF@SEPLN (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Peters</surname>
            ,
            <given-names>M.E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Neumann</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Iyyer</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gardner</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Clark</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zettlemoyer</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Deep contextualized word representations</article-title>
          .
          <source>CoRR abs/1802</source>
          .05365 (
          <year>2018</year>
          ), http://arxiv.org/abs/
          <year>1802</year>
          .05365
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <article-title>Plaza-del-</article-title>
          <string-name>
            <surname>Arco</surname>
            ,
            <given-names>F.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jimenez-Zafra</surname>
            ,
            <given-names>S.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Montejo-Raez</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Molina-Gonzalez</surname>
            ,
            <given-names>M.D.</given-names>
          </string-name>
          ,
          <article-title>Uren~a-</article-title>
          <string-name>
            <surname>Lopez</surname>
            ,
            <given-names>L.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mart</surname>
          </string-name>
          n-Valdivia, M.T.:
          <article-title>Overview of the EmoEvalEs task on emotion detection for Spanish at IberLEF 2021</article-title>
          .
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>67</volume>
          (
          <issue>0</issue>
          ) (
          <year>2021</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <article-title>Plaza-del-</article-title>
          <string-name>
            <surname>Arco</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Strapparava</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Uren</surname>
            <given-names>~</given-names>
          </string-name>
          <article-title>a-</article-title>
          <string-name>
            <surname>Lopez</surname>
            ,
            <given-names>L.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Martin-Valdivia</surname>
          </string-name>
          , M.T.:
          <article-title>EmoEvent: A Multilingual Emotion Corpus based on di erent Events</article-title>
          .
          <source>In: Proceedings of the 12th Language Resources and Evaluation Conference</source>
          . pp.
          <volume>1492</volume>
          {
          <fpage>1498</fpage>
          .
          <string-name>
            <surname>European Language Resources Association</surname>
          </string-name>
          , Marseille, France (May
          <year>2020</year>
          ), https://www.aclweb.org/anthology/2020.lrec-
          <volume>1</volume>
          .
          <fpage>186</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Radford</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Narasimhan</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Improving language understanding by generative pre-training (</article-title>
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>Tang</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Qin</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Feng</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
          </string-name>
          , T.:
          <article-title>E ective LSTMs for target-dependent sentiment classi cation</article-title>
          .
          <source>In: Proceedings of COLING</source>
          <year>2016</year>
          ,
          <source>the 26th International Conference on Computational Linguistics: Technical Papers</source>
          . pp.
          <volume>3298</volume>
          {
          <fpage>3307</fpage>
          .
          <string-name>
            <surname>The</surname>
            <given-names>COLING</given-names>
          </string-name>
          2016
          <string-name>
            <given-names>Organizing</given-names>
            <surname>Committee</surname>
          </string-name>
          , Osaka,
          <source>Japan (Dec</source>
          <year>2016</year>
          ), https://www.aclweb. org/anthology/C16-1311
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Wagner</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Arora</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cortes</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Barman</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bogdanova</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Foster</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tounsi</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>DCU: Aspect-based polarity classi cation for SemEval task 4</article-title>
          .
          <source>In: Proceedings of the 8th International Workshop on Semantic Evaluation (SemEval</source>
          <year>2014</year>
          ). pp.
          <volume>223</volume>
          {
          <fpage>229</fpage>
          . Association for Computational Linguistics, Dublin, Ireland (Aug
          <year>2014</year>
          ). https://doi.org/10.3115/v1/
          <fpage>S14</fpage>
          -2036, https://www.aclweb. org/anthology/S14-2036
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liakata</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zubiaga</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Procter</surname>
          </string-name>
          , R.:
          <article-title>TDParse: Multi-target-speci c sentiment recognition on Twitter</article-title>
          .
          <source>In: Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics: Volume</source>
          <volume>1</volume>
          ,
          <string-name>
            <given-names>Long</given-names>
            <surname>Papers</surname>
          </string-name>
          . pp.
          <volume>483</volume>
          {
          <fpage>493</fpage>
          . Association for Computational Linguistics, Valencia,
          <source>Spain (Apr</source>
          <year>2017</year>
          ), https://www.aclweb.org/anthology/E17-1046
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhu</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zhao</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Attention-based LSTM for aspect-level sentiment classi cation</article-title>
          .
          <source>In: Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing</source>
          . pp.
          <volume>606</volume>
          {
          <fpage>615</fpage>
          . Association for Computational Linguistics, Austin, Texas (Nov
          <year>2016</year>
          ). https://doi.org/10.18653/v1/
          <fpage>D16</fpage>
          - 1058, https://www.aclweb.org/anthology/D16-1058
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>