<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>UMUTeam at EmoEvalEs 2021: Emotion Analysis for Spanish based on Explainable Linguistic Features and Transformers</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Antonio G</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ricardo Colomo-Palacios</string-name>
          <email>ricardo.colomo-palacios@hiof.no</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Facultad de Informatica, Universidad de Murcia, Campus de Espinardo</institution>
          ,
          <addr-line>30100</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Faculty of Computer Sciences, stfold University College</institution>
          ,
          <addr-line>Halden</addr-line>
          ,
          <country country="NO">Norway</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Emotion Analysis extends the idea of Sentiment Analysis by shifting from plain positive or negative sentiments to a rich variety of emotions to get better understanding of the users' thoughts and appraisals. The move from Sentiment Analysis to Emotion Analysis requires, however, better feature engineering techniques when it comes to capturing complex language phenomena, which have to do with gurative language and the way of expressing oneself. In this manuscript we detail the participation of the UMUTeam in EmoEvalEs'2021 shared task from IberLEF, concerning the identi cation of emotions in Spanish. Our proposal is grounded on the combination of explainable linguistic features and state-of-the-art transformers based on the Spanish version of BERT. We achieved the 6th position in the o cial leader board with an accuracy of 68.5990%, only 4.1667% below the best result. In addition, we apply model agnostic techniques for explainable arti cial intelligence to achieve insights from the linguistic features. We observed a correlation between psycho-linguistic processes and perceptual feel with the emotions evaluated and, speci cally, with documents labelled as sadness.</p>
      </abstract>
      <kwd-group>
        <kwd>Emotion Analysis</kwd>
        <kwd>Feature Engineering</kwd>
        <kwd>Natural Language Processing</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>Emotion Analysis (EA) is a Natural Language Processing (NLP) task related
to Sentiment Analysis (SA), Document Classi cation (DC) and Information
Retrieval (IR), whose objective is the identi cation of emotions from a piece of
text [21]. Standard SA, on the other hand, is focused on determining whether
a document is positive, neutral, or negative. EA insights, therefore, are useful
for creating better recommender systems that adapt better to the mood of the
users [21]. Moreover, the oversimpli cation of SA could be misleading in some
scenarios. For example, while analysing online reviews of movies, EA might
identify as sadness the emotions that arouse in people from the lm La vita e bella;
however, these reviews can be wrongly classi ed as negative from conventional
SA approaches because sadness and negative feelings are related in some way
[9].</p>
      <p>In this manuscript we describe the participation of the UMUTeam in the
shared task EmoEvalEs 2021 [15] proposed at Iberian Languages Evaluation
Forum (IberLEF) [11]. This task is focused on the classi cation of emotions
in micro-blogging posts, which is challenging due to the absence of contextual
clues such as voice modulation or facial expressions. Speci cally, this task aims
to distinguish among the following emotions: Anger, Disgust, Fear, Joy, Sadness,
Surprise and Others.</p>
      <p>One of our objectives for participating in this task is the evaluation of a set
of linguistic characteristics extracted with the tool UMUTextStats [4, 5] of which
it is part a doctoral thesis from a team member. It is worth mentioning that we
participated with an previous version of this tool on TASS 2020 shared task [6],
in which a similar EA subtask was proposed. However, for this task we present a
major revision of the linguistic features and new forms of combining them with
state-of-the-art transformers.</p>
      <p>This manuscript is organised as follows: First, Section 2 provides background
information regarding EA. Next, Section 3 describes brie y the corpus that was
made available by the organisers of the shared task. The methodology is depicted
in Section 4. Next, Section 5 contains the results achieved by our team and the
comparison with the rest of the participants. In addition, an interpretation of
the features is presented. Finally, the conclusions and promising future research
directions are shown in Section 6.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Background information</title>
      <p>Emotion categorisation is a challenging task. On the one hand, there are several
emotion classi cations [7], such as six Ekman's basic emotions [3], Plutchik's
Wheel of Emotions [17] or Russel's Circumplex Model [20]. On the other, the
detection of emotions is subtle to distinguish as several emotions can be present
at the same time. Also, there are not too many studies and resources in Spanish
focused on this task. Nevertheless, recent shared tasks are focusing on Spanish
EA as TASS 2020 [24], which includes a subtask based on six Ekman's basic
emotions in Spanish tweets. One of the approaches to address the lack of datasets
in Spanish for EA was carried out in [16], in which the authors presented a
dataset of tweets compiled in April 2019 annotated based on the six Ekman's
basic emotions plus an extra emotion for neutral and others. Another recent
work is [1], in which the authors apply EA to social media by incorporating to
their pipeline a ective lexical resources such as SEL [22], iSOL [10], and EmoLex
[8]. The experiments performed in this work indicate that the usage of linguistic
features and sentiment lexicons are advantageous for conducting EA. In the same
line, the usage of linguistic features have proven e ective in other related tasks
such as satire identi cation [14], in which the authors employ Linguistic Inquiry
Word Count (LIWC) [23] for distinguish among satiric and non-satiric texts from
European Spanish and Mexican Spanish tweets.
3</p>
    </sec>
    <sec id="sec-3">
      <title>Corpus</title>
      <p>According to the organisers of the task, the EmoEvalEs' dataset consisted in
tweets from April 2019 based on di erent events. The tweets were pre-processed
to replace hashtags and mentions with some tokens to hinder the automatic
classi cation task. The dataset was distributed in three splits: train, development,
and testing. Table 1 depicts the distribution of the corpus. As we can observe,
many of the tweets could not be labelled with one of the sentiments and they
were rated as others. This fact gives an idea of the di culty of the task, even
for human annotators. The emotions with more number of instances are joy,
and sadness that are, from our point of view, the most generic and polarised
emotions. On contrast, fear and disgust emotions are underrepresented in the
dataset. It may be that these emotions are di cult to categorise, or that people
do not express those emotions on public social networks.
This section describes the feature sets employed for solving this task, the neural
networks evaluated, and the hyperparameter optimisation stage carried out.</p>
      <p>
        Regarding the features employed, our proposal is grounded on linguistic
features in combination with state-of-the-art transformers [25]. During our
experimentation, we also evaluated word and sentence embeddings from pre-trained
Spanish models. For the linguistic features (LF ) we use UMUTextStats [4, 5].
This tool is inspired in LIWC [23] but designed from scratch the Spanish
language. UMUTextStats takes into account more than 350 linguistic features
categorised as follows: (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) phonetics, which handles techniques such as word
elongation; (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) morphosyntax, that includes a ne-grained Part-of-Speech tags
extracted from Stanza [18] and custom lexicons; (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ) correction and style, that
captures di erent stylistic and correction patterns used during writing; (
        <xref ref-type="bibr" rid="ref4">4</xref>
        )
semantics, that captures linguistic phenomena such as onomatopoeia, euphemism,
dynamism, or synecdoc; (
        <xref ref-type="bibr" rid="ref5">5</xref>
        ) pragmatics, that includes gurative language
phenomena [13], discourse markers and courtesy forms; (
        <xref ref-type="bibr" rid="ref6">6</xref>
        ) stylometry, including
several corpora statistics such as Type-token ratio (TTR) and punctuation
symbols; (
        <xref ref-type="bibr" rid="ref7">7</xref>
        ) lexical, that includes a wide variety of topics, including locations,
organisations, animals, weapons, food, religion, or health among others; (
        <xref ref-type="bibr" rid="ref8">8</xref>
        )
psycholinguistic processes, that includes positive and negative expressions; (9) register,
that includes the usage of informal speech, colloquialisms, or SMS language; and
(10) social media, that captures jargon used in social networks. For the
transformers we use the Spanish version of BERT, also known as BETO [2]. To obtain
these vectors, we evaluated two methods, that we called BE and BF respectively,
both extracting the [CLS] token in a similar way as is detailed at [19] and using
HuggingFace (v4.4.2). The key di erence is that for BE we obtained the vectors
from BETO directly, whereas for BF we rst ne-tuned BETO with the
EmoEvalEs dataset. Both BE and BF are xed-vectors of 768 items per document.
In addition to the transformers, we also evaluated neural networks with word
and sentences embeddings from fastText, word2vec, and gloVe. We refer to these
feature set as WE for the word embeddings and SE for the sentence embeddings.
      </p>
      <p>Each feature set (LF, SE, BF, and WE ) was trained separately and in
combination using the functional API of Keras. For the xed-sentence vectors we rely
on multi-layer perceptrons but for WE we also evaluated a convolutional and
two bidirectional recurrent neural networks, based on Long-Short Term Memory
(BiLSTM) and Gated Recurrent Unit (BiGRU), that have provided good results
in the past for conducting SA tasks [12].</p>
      <p>
        The next step in our pipeline consisted in a hyperparameter optimisation.
For this, we evaluate a total of 110 neural models per feature set (in
isolation or combined). The best model was selected using the weighted F1 score.
Most of the neural networks evaluated consisted in shallow multilayer
perceptrons (MLP) with one or two hidden layers and with both hidden layers having
the same number of neurons (
        <xref ref-type="bibr" rid="ref8">8, 16, 48, 64, 128, 256</xref>
        ). We also evaluated deep
neural networks with a number of hidden layers between 3 and 8, with a di
erent number of neurons per hidden layer organised in di erent shapes. For the
rest of the hyper-parameters, we evaluated di erent dropout rates, several
activation functions, and di erent learning rates. The source code is available at
https://github.com/Smolky/emoevales-2021.
      </p>
      <p>https://huggingface.co/sentence-transformers/bert-base-nli-cls-token
https://huggingface.co/
https://github.com/dccuchile/spanish-word-embeddings</p>
      <p>Table 2 depicts the results of the hyperparameter optimisation stage for each
feature set separately and in combination. For the sake of simplicity, we have
included only the combinations with LF. Regarding the feature sets separately,
we can observe that the best results are obtained with shallow neural networks,
with 2 hidden layers (except for SE) with brick shape. The number of neurons is
always less than the number of parameters, resulting in 256 neurons for LF, 128
for SE, and 512 for BE and BF. All neural networks achieved their best results
with dropout for the features in isolation. The learning rate varies from 0.001 for
LF and BE to 0.01 for SE and BF. Out of the activation functions, relu achieves
better results for LF, SE, and BE whereas tanh achieves better results for BF.
When we observe the features combined in pairs, only the combination of LF
with BE requires a complex deep neural network to achieve their best result,
with 4 hidden layers and 512 neurons stacked in a diamond shape. However,
when combining LF with BF, the best result is achieved with a simpler model
composed by two hidden layers of 128 neurons each. When combined in groups
of three, the combination of LF, SE, and BE requires also a deep neural network
composed of four hidden layers (as the combination of LF with BE) but with
1024 neurons. However, the combination of LF, SE, and BF resulted in a simpler
model of two hidden layers with 48 neurons each. A similar architecture can be
found when combining LF, SE, BE, and BF. In this case, the network also results
in a very simpler model with only one hidden layer of 48 neurons, a dropout
of 0.2, and a learning rate of 0.01 with a sigmoid as activation function. The
simplicity of the networks in which BF is present can be explained because the
weights of BF have been trained with the EmoEvalEs dataset, so the embeddings
have already been grouped based on the emotions within the latent space.</p>
    </sec>
    <sec id="sec-4">
      <title>Results</title>
      <p>Participants were required to submit a maximum of three runs that are ranked
by macro average F1-score but also by accuracy and the macro-averaged versions
of Precision and Recall. The organisers of the task allowed the participants to
send their runs in two separated time slots: A development phase, in which the
participants could evaluate their results with the development dataset, and the
o cial one, against the test split. Due to lack of time, we were able to send only
a run during the development phase that achieved an accuracy of 70.8531% and
a macro averaged F1-Score of 69.9542%, reaching the second position in a total
of six participants.</p>
      <p>For the o cial competition our rst run consisted in an ensemble of the best
model for each feature set: LF, SE, BE, and BF. We exclude WE because it
requires a large amount of time for training and the results does not outperform
the models based on xed-length vectors. This ensemble model decides the nal
output with an averaged version of the mode. For that, we store the results of
each model with the validation set in order to decide its weight for the nal
decision. We achieve an 68.599% accuracy with this run. The macro F1-Score is
66.8407%, the precision is 67.2546%, and the recall is 68.5990%. For our second
run, we evaluate another form of ensemble based on the softmax layer of each
neural network. We use the probabilities of each neural network to train an
extra ensemble. This run achieves worst result than the previous ensemble with
an accuracy of 68.2971%. Our last submission consisted in a MLP perceptron
trained with two inputs LF, and BF, as we want to compare the results of
methods non based on ensembles. We achieve an accuracy of 66.7874%.</p>
      <p>The o cial results are depicted in Table 3. We achieve the 6th position in the
o cial leader board with an accuracy of 68.5990%, a macro average precision of
67.2546%, a macro recall of 68.5990%, and a macro F1-score of 66.8407%. The
best result is achieved by fyinh, with an macro F1-score of 71.7028%, followed
by fyinh with an macro F1-score of 71.1373%. We can observe that all runs and
participants achieve competitive results. On the one hand, the major accuracy
di erence is only of 10.9903% between the best and worst result. On the other
hand, the relation regarding the macro precision and macro recall is similar
among all the participants. It is worth noting that we set the main metric for
the hyper-parameter optimisation to the weighted f1-score but nally, macro
f1score was the o cial score. It is possible, therefore, that we could achieve better
results with a better strategy.</p>
      <p>We include the normalised confusion matrix of the best model, a ensemble
that combines LF, SE, BE, and BF using the weighted mode, with the validation
set (see Figure 1). We can observe that anger is predicted correctly most of the
times, and the wrong classi cations are about the others class. Emotions of
disgust are classi ed wrongly as anger, followed by others and fear. Only a 12%
of documents labelled as disgust are correctly classi ed. Documents labelled as
fear by the annotators are correctly classi ed the 67%, but sometimes they are
wrongly classi ed as anger, disgust, and others. It is worth noting that both
classes, fear and disgust are the labels with less instances in the corpus and that
our proposal is especially confused with the disgust class. For the class joy, our
system classi es it correctly the 61%, labelling as others a 32%. The majority
class others is classi ed correctly the 80%, but the 12% is wrongly classi ed as
joy. Note that these were the classes with large number of instances. Sadness
is correctly classi ed the 67%. Finally, for the documents labelled as surprise,
our system is able to classify the 37%, but a 43% of the times are classi ed as
others, an 11% as anger, and a 9% as joy. The strong points of our proposal are
that there are not so many wrongly classi cations as opposite emotions, as it
could be labelling sadness as joy or vice-versa. However, our proposal is confused
between anger and disgust and it achieves a low recall on the class surprise.</p>
      <p>In order to provide some understanding of the linguistic features, we obtain
the top ten discriminatory linguistic features per class (see Figure 2) and we
generate a polar chart for each linguistic category and emotion (see Figure 3).
Note that in both charts we exclude intentionally those tweets labelled as others.</p>
      <p>As it was expected, the features related to sad emotions are strong
discriminatory for the sadness label, but also has an strong impact on disgust. In a similar
manner, anger label is also related with the psycho-linguistic process anger, but
also with disgust. As a personal opinion, anger and disgusts are the emotions in
which it is more di cult to di erentiate. Another correlation is perceptual feel,
which has a strong correlation with sadness. In the same line, negative process
is also related to di erent emotions such as anger, disgust, fear, and sadness,
but also is relevant for documents labelled as surprise. It draws our attention
that the token º has a strong correlation for documents labelled as sadness and
surprise. We manually checked which tweets contains that sign and the majority
are related to the sports events, such as La Liga and ChampionsLeague. They
appear to discuss about results by means of ordinal numbers. It can also be
ltcauA ssdauoidarstpnnghfregeeujiossearseyssrrt a1671264n1241%%%g%%%%er d11i16021s12%%%%%g%%ust 6060020f7e%%%%%%%arPredicte6100193d12jo%%%%%%%y o3111428t2195300h%%%%%%%ers sa6100201d7%%%%%%n%ess 33001027%%%%%%%
Fig. 1. Confusion matrix with the validation split with an ensemble based on the
weighted mode of LF, SE, BE, and BF
observed that tweets with fewer words correspond mostly to tweets labelled as
fear and joy.</p>
      <p>anger
disgust fear joy
sadness
surprise
psycholinguistic processes</p>
      <p>negative sad
psycholinguistic processes</p>
      <p>negative general
psycholinguistic processes</p>
      <p>negative
lexical social perceptual feel
psycholinguistic processes</p>
      <p>positive
psycholinguistic processes</p>
      <p>negative anger
psycholinguistic processes
stylometry punctuation symbols</p>
      <p>numero sign
psycholinguistic processes</p>
      <p>positive general
stylometry corpus words count
0%
25%
50%
75%
100%</p>
      <p>Regarding each linguistic feature category (see Figure 3), the major di
erence among emotions appears in the semantics category, that it is the one that
includes positive, and negative emotions. Regarding phonetics, that include
features such as word elongation to add emphasis, sadness is the emotion that
makes less use of this linguistic device. Regarding correction and style, fear is
the emotion in which most stylistic errors are detected. Regarding lexical and
topics, there is a wide heterogeneity among the emotions, ordered from major
to minor use of topics by surprise, joy, fear, disgust, anger, and sadness. This
fact suggest that people describes the cause of their emotions to explain which
causes its surprise or joy, but they are less likely to explain why they are sad or
angry.
6</p>
    </sec>
    <sec id="sec-5">
      <title>Conclusions</title>
      <p>Here we have described the participation of the UMUTeam at EmoEvalEs 2021
shared task regarding EA in Spanish. As commented earlier, this task has been
anger disgust fear joy
sadness</p>
      <p>surprise
an opportunity for us to evaluate our methods in real scenarios and we considered
that we achieved competitive results but with room for improvement. From the
point of view of explainable arti cial intelligence, we have shown the potential
of the linguistic features to provide model agnostic methods for explainability.</p>
      <p>As promising research directions we suggest to continue with the
interpretability of the neural network models and features. In this sense, we propose to
nd the correlations between the linguistic features and embeddings in order
to determine in which cases they are complementary and in which not. Another
promising direction is to provide contextual features to EA, in order to track how
sentiments and emotions are changing on online conversations such as threads
on Twitter.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This work was supported by the Spanish National Research Agency (AEI)
through project LaTe4PSP (PID2019-107652RB-I00/AEI/10.13039/501100011033).
In addition, Jose Antonio Garc a-D az has been supported by Banco Santander
and University of Murcia through the industrial doctorate programme.
9. Mokryn, O., Bodo , D., Bader, N., Albo, Y., Lanir, J.: Sharing emotions:
determining lms' evoked emotional experience from their online reviews. Information
Retrieval Journal 23, 475{501 (2020)
10. Molina-Gonzalez, M.D., Mart nez-Camara, E., Mart n-Valdivia, M.T.,
PereaOrtega, J.M.: Semantic orientation for polarity classi cation in spanish reviews.</p>
      <p>
        Expert Systems with Applications 40(18), 7250{7257 (2013)
11. Montes, M., Rosso, P., Gonzalo, J., Aragon, E., Agerri, R., Alvarez-Carmona,
M.A., Alvarez Mellado, E., Carrillo-de Albornoz, J., Chiruzzo, L., Freitas, L.,
Gomez Adorno, H., Gutierrez, Y., Jimenez Zafra, S.M., Lima, S., Plaza-de Arco,
F.M., Taule, M.: Proceedings of the iberian languages evaluation forum (iberlef
2021). In: CEUR workshop (2021)
12. Paredes-Valverde, M.A., Colomo-Palacios, R., Salas-Zarate, M.d.P.,
ValenciaGarc a, R.: Sentiment analysis in spanish for improvement of products and services:
a deep learning approach. Scienti c Programming 2017 (2017)
13. del Pilar Salas-Zarate, M., Alor-Hernandez, G., Sanchez-Cervantes, J.L.,
ParedesValverde, M.A., Garc a-Alcaraz, J.L., Valencia-Garc a, R.: Review of
english literature on gurative language applied to social networks. Knowl.
Inf. Syst. 62(
        <xref ref-type="bibr" rid="ref6">6</xref>
        ), 2105{2137 (2020). https://doi.org/10.1007/s10115-019-01425-3,
https://doi.org/10.1007/s10115-019-01425-3
14. del Pilar Salas-Zarate, M., Paredes-Valverde, M.A., Rodr guez-Garc a,
M.A., Valencia-Garc a, R., Alor-Hernandez, G.: Automatic detection
of satire in twitter: A psycholinguistic-based approach. Knowl. Based
Syst. 128, 20{33 (2017). https://doi.org/10.1016/j.knosys.2017.04.009,
https://doi.org/10.1016/j.knosys.2017.04.009
15. Plaza-del-Arco, F.M., Jimenez-Zafra, S.M., Montejo-Raez, A., Molina-Gonzalez,
M.D., Uren~a-Lopez, L.A., Mart n-Valdivia, M.T.: Overview of the EmoEvalEs task
on emotion detection for Spanish at IberLEF 2021. Procesamiento del Lenguaje
Natural 67(0) (2021)
16. Plaza-del-Arco, F., Strapparava, C., Uren~a-Lopez, L.A., Mart n-Valdivia, M.T.:
EmoEvent: A Multilingual Emotion Corpus based on di erent Events. In:
Proceedings of the 12th Language Resources and Evaluation Conference. pp. 1492{
1498. European Language Resources Association, Marseille, France (May 2020),
https://www.aclweb.org/anthology/2020.lrec-1.186
17. Plutchik, R., Kellerman, H.: Theories of emotion, vol. 1. Academic Press (2013)
18. Qi, P., Zhang, Y., Zhang, Y., Bolton, J., Manning, C.D.: Stanza: A python
natural language processing toolkit for many human languages. arXiv preprint
arXiv:2003.07082 (2020)
19. Reimers, N., Gurevych, I.: Sentence-bert: Sentence embeddings using siamese
bertnetworks. CoRR abs/1908.10084 (2019), http://arxiv.org/abs/1908.10084
20. Russell, J.A.: A circumplex model of a ect. Journal of personality and social
psychology 39(
        <xref ref-type="bibr" rid="ref6">6</xref>
        ), 1161 (1980)
21. Sailunaz, K., Dhaliwal, M., Rokne, J., Alhajj, R.: Emotion detection from text and
speech: a survey. Social Network Analysis and Mining 8(
        <xref ref-type="bibr" rid="ref1">1</xref>
        ), 1{26 (2018)
22. Sidorov, G., Miranda-Jimenez, S., Viveros-Jimenez, F., Gelbukh, A.,
CastroSanchez, N., Velasquez, F., D az-Rangel, I., Suarez-Guerra, S., Trevino, A.,
Gordon, J.: Empirical study of machine learning based approach for opinion mining
in tweets. In: Mexican international conference on Arti cial intelligence. pp. 1{14.
      </p>
      <p>
        Springer (2012)
23. Tausczik, Y.R., Pennebaker, J.W.: The psychological meaning of words: Liwc and
computerized text analysis methods. Journal of language and social psychology
29(
        <xref ref-type="bibr" rid="ref1">1</xref>
        ), 24{54 (2010)
24. Vega, M.G., D az-Galiano, M.C., Cumbreras, M.A.G., del Arco, F.M.P.,
MontejoRaez, A., Zafra, S.M.J., Camara, E.M., Aguilar, C.A., Cabezudo, M.A.S., Chiruzzo,
L., Moctezuma, D.: Overview of TASS 2020: Introducing emotion detection. In:
Cumbreras, M.A.G., Gonzalo, J., Camara, E.M., Mart nez-Unanue, R., Rosso, P.,
Zafra, S.M.J., Zambrano, J.A.O., Miranda, A., Zamorano, J.P., Gutierrez, Y.,
Rosa, A., Montes-y-Gomez, M., Vega, M.G. (eds.) Proceedings of the Iberian
Languages Evaluation Forum (IberLEF 2020) co-located with 36th Conference
of the Spanish Society for Natural Language Processing (SEPLN 2020), Malaga,
Spain, September 23th, 2020. CEUR Workshop Proceedings, vol. 2664, pp. 163{
170. CEUR-WS.org (2020), http://ceur-ws.org/Vol-2664/tass overview.pdf
25. Wolf, T., Chaumond, J., Debut, L., Sanh, V., Delangue, C., Moi, A., Cistac, P.,
Funtowicz, M., Davison, J., Shleifer, S., et al.: Transformers: State-of-the-art
natural language processing. In: Proceedings of the 2020 Conference on Empirical
Methods in Natural Language Processing: System Demonstrations. pp. 38{45 (2020)
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Plaza-del Arco</surname>
            ,
            <given-names>F.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mart</surname>
            n-Valdivia,
            <given-names>M.T.</given-names>
          </string-name>
          ,
          <article-title>Uren~a-</article-title>
          <string-name>
            <surname>Lopez</surname>
            ,
            <given-names>L.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mitkov</surname>
          </string-name>
          , R.:
          <article-title>Improved emotion recognition in spanish social media through incorporation of lexical knowledge</article-title>
          .
          <source>Future Generation Computer Systems</source>
          <volume>110</volume>
          ,
          <fpage>1000</fpage>
          {
          <fpage>1008</fpage>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2. Can~ete, J.,
          <string-name>
            <surname>Chaperon</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fuentes</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Perez</surname>
          </string-name>
          , J.:
          <article-title>Spanish pre-trained bert model and evaluation data</article-title>
          .
          <source>PML4DC at ICLR</source>
          <year>2020</year>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Ekman</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Lie catching and microexpressions</article-title>
          .
          <source>The philosophy of deception 1(2)</source>
          ,
          <volume>5</volume>
          (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Garc</surname>
          </string-name>
          a
          <article-title>-D az</article-title>
          ,
          <string-name>
            <given-names>J.A.</given-names>
            ,
            <surname>Canovas-Garc</surname>
          </string-name>
          <string-name>
            <given-names>a</given-names>
            , M.,
            <surname>Valencia-Garc</surname>
          </string-name>
          <string-name>
            <surname>a</surname>
          </string-name>
          , R.:
          <article-title>Ontology-driven aspect-based sentiment analysis classi cation: An infodemiological case study regarding infectious diseases in latin america</article-title>
          .
          <source>Future Generation Computer Systems</source>
          <volume>112</volume>
          ,
          <fpage>614</fpage>
          {
          <fpage>657</fpage>
          (
          <year>2020</year>
          ). https://doi.org/10.1016/j.future.
          <year>2020</year>
          .
          <volume>06</volume>
          .019
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Garc</surname>
          </string-name>
          a
          <article-title>-D az</article-title>
          ,
          <string-name>
            <given-names>J.A.</given-names>
            ,
            <surname>Canovas-Garc</surname>
          </string-name>
          <string-name>
            <given-names>a</given-names>
            , M.,
            <surname>Colomo-Palacios</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Valencia-Garc</surname>
          </string-name>
          <string-name>
            <surname>a</surname>
          </string-name>
          , R.:
          <article-title>Detecting misogyny in spanish tweets. an approach based on linguistics features and word embeddings</article-title>
          .
          <source>Future Generation Computer Systems</source>
          <volume>114</volume>
          ,
          <fpage>506</fpage>
          {
          <fpage>518</fpage>
          (
          <year>2021</year>
          ). https://doi.org/10.1016/j.future.
          <year>2020</year>
          .
          <volume>08</volume>
          .032, http://www.sciencedirect.com/science/article/pii/S0167739X20301928
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Garc</surname>
          </string-name>
          a
          <article-title>-D az</article-title>
          ,
          <string-name>
            <given-names>J.A.</given-names>
            ,
            <surname>Almela</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Valencia-Garc</surname>
          </string-name>
          <string-name>
            <surname>a</surname>
          </string-name>
          , R.: Umuteam at tass 2020:
          <article-title>Combining linguistic features and machine-learning models for sentiment classi cation</article-title>
          .
          <source>Proceedings of TASS</source>
          (
          <year>2020</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Klinger</surname>
            ,
            <given-names>R.:</given-names>
          </string-name>
          <article-title>A survey on sentiment and emotion analysis for computational literary studies</article-title>
          . CoRR abs/
          <year>1808</year>
          .03137 (
          <year>2018</year>
          ), http://arxiv.org/abs/
          <year>1808</year>
          .03137
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Mohammad</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Turney</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>Emotions evoked by common words and phrases: Using mechanical turk to create an emotion lexicon</article-title>
          .
          <source>In: Proceedings of the NAACL HLT</source>
          <year>2010</year>
          <article-title>workshop on computational approaches to analysis and generation of emotion in text</article-title>
          . pp.
          <volume>26</volume>
          {
          <issue>34</issue>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>