<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">UMUTeam at TASS 2020: Combining Linguistic Features and Machine-learning Models for Sentiment Classification</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">José</forename><forename type="middle">Antonio</forename><surname>García-Díaz</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Facultad de Informática</orgName>
								<orgName type="institution">Universidad de Murcia</orgName>
								<address>
									<addrLine>Campus de Espinardo</addrLine>
									<postCode>30100</postCode>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Ángela</forename><surname>Almela</surname></persName>
							<email>angelalm@um.es</email>
							<affiliation key="aff1">
								<orgName type="department">Facultad de Letras</orgName>
								<orgName type="institution">Universidad de Murcia</orgName>
								<address>
									<addrLine>Campus de La Merced</addrLine>
									<postCode>30001</postCode>
									<settlement>Murcia</settlement>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Rafael</forename><surname>Valencia-García</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Facultad de Informática</orgName>
								<orgName type="institution">Universidad de Murcia</orgName>
								<address>
									<addrLine>Campus de Espinardo</addrLine>
									<postCode>30100</postCode>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Valencia</forename><surname>-García</surname></persName>
							<email>valencia@um.es</email>
						</author>
						<title level="a" type="main">UMUTeam at TASS 2020: Combining Linguistic Features and Machine-learning Models for Sentiment Classification</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">DCA7A41B015EA4A45DB6C3F8E962B508</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T04:20+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Sentiment Analysis</term>
					<term>Supervised learning</term>
					<term>Deep Learning</term>
					<term>Machine learning</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper describes the participation of the UMUTeam at the TASS'2020 Workshop on Sentiment Analysis, in which two tracks were proposed. The first track consists in the classification of tweets according to general sentiments of tweets written in several Spanish varieties, whereas the second task consists in a fine-grained distinction between the six basic emotions described by Ekman (2009). Our proposal is based on the usage of linguistic features alone or in combination with word-embeddings. Specifically, we test Convolutional Neural Networks and Support Vector Machines with sentence embeddings. Although our proposal did not achieve the best results, we obtained the best precision rate regarding emotion detection (Task 2) and competitive results with respect to the general sentiment classification in which tweets written in different varieties of Spanish were mixed. We consider that our proposal, despite its limitations, provides substantial benefits such as the interpretability of the results.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>This paper describes the participation of the UMUTeam in the TASS'2020 Workshop on Sentiment Analysis (SA). SA is a trending task of Natural Language Processing (NLP) which consists in the extraction and classification of subjective opinions in texts written in natural language.</p><p>Two tasks were proposed to the participants. The first one involved extracting the subjectivity polarity from tweets written in different varieties of Spanish, including European Spanish and different varieties from Latin America. This task was divided into two sub-tasks: (1) monolingual classification, in which several training and testing splits from different Spanish varieties were considered separately: Spain (es), Costa Rica (cr), Uruguay (uy), and Mexico (mx); and (2) multi-variant classification, in which a testing dataset consisting of mixed tweets from the above Spanish varieties were mixed. In both sub-tasks the participants were required to determine if tweets express positive, negative or neutral sentiments. The second task consisted in a multiclass classification of emotions based on six of the basic emotions proposed by <ref type="bibr" target="#b0">Ekman (2009)</ref> [1]: anger, disgust, fear, joy, sadness and surprise along with others in order to classify those tweets which do not agree with the aforementioned labels. All datasets, regardless of the task, were unbalanced: Task 1 contained more tweets labelled as neutral and Task 2 contained more tweets labelled as others.</p><p>We participated both in Task 1 and Task 2. In a nutshell, our proposal consisted in testing the reliability of linguistic features alone or in combination with well-known machine-learning and deep-learning models. Linguistic features, in comparison with statistical approaches, ease the interpretability of the results because they are conceptually higher-level features than words. With our participation, we attempted to answer the following research questions:</p><p>• RQ1. Is the usage of linguistic features enough to compete with state-of-the-art approaches based on statistical methods?</p><p>• RQ2. Can linguistic features be combined with statistical methods in order to improve the accuracy of the results while keeping interpretability?</p><p>• RQ3. Is the reliability of linguistic features affected by different cultural background in the same language?</p><p>To answer these research questions, we extracted linguistic features from the datasets by using a self-developed tool designed from scratch for the Spanish language and which is part of the first author's doctoral thesis. These linguistic features were evaluated separately or in combination with word-embeddings with a Convolutional Neural Network (CNN) and with sentence-embeddings with Support Vector Machines (SVM).</p><p>The rest of the paper is organised as follows. In Section 2, the datasets provided to the participants are described, whereas Section 3 describes the materials and methods used in our proposal. Then, Section 4 shows the results from each model and task along with an analysis of the results. Finally, Section 5 summarises the results according to the research questions and describes further improvement actions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Corpus description</head><p>The datasets involved in these tasks were provided by the organisers of the TASS workshop <ref type="bibr" target="#b1">[2]</ref>. The corpus was compiled from Twitter in April 2019 and it contains topics from different domains, such as politics, entertainment, catastrophes, and global events, among others. The corpus contains approximately 16k tweets for Task 1, and 8.4k tweets for Task 2 <ref type="bibr" target="#b2">[3]</ref>. For each task, the corpus was released as three datasets, namely, training, development, and testing. Table <ref type="table" target="#tab_0">1</ref> shows the distribution of the datasets for each task. Note that the size of the training and development datasets of Subtask 1.2 is 0 because participants were expected to use the training and development datasets of Subtask 1.1.</p><p>It is worth noting that hashtags and mentions were replaced by the tokens HASHTAG and @USER in order to prevent that automatic classifiers could overfit their results based on wrong assumptions. Another significant fact was that the label Neutral also referred to those tweets where no sentiment was assigned during the manual classification of the corpus. This is a novelty with respect to previous editions of the TASS Workshop, in which tweets had been classified as neutral and none separately.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Materials and methods</head><p>In this section, we describe (1) the preprocessing techniques applied to the corpus according to our proposal (see Section 3.1), (2) the linguistic features extracted (see Section 3.2), and (3) the machine-learning models used for training the sentiment classifiers (see Section 3.3).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Preprocessing stage</head><p>Our preprocessing stage involved (1) encoding each letter to its lowercase form, (2) fixing misspelling and typos by using the Aspell library<ref type="foot" target="#foot_0">1</ref> , (3) contracting white-space characters, such as spaces, tabs or new lines, and (4) removing expressive lengthening (the intentional elongation of letters in a word to emphasise it). The normalised version of each tweet is used as input for extracting those linguistic features that are dictionary based. However, we also kept the original version of the tweet to extract certain features regarding the usage of uppercase letters (which may be indicative of shouting or emphasis) or to identify the number of misspellings among other features.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Linguistic feature extraction</head><p>For the extraction of linguistic features we used UMUTextStats <ref type="bibr" target="#b3">[4]</ref>, a self-developed linguistic tool for text analysis that is based on Linguistic Inquiry and Word Count (LIWC) <ref type="bibr" target="#b4">[5]</ref>. LIWC is an analysis tool that counts words which belong to pre-established categories based on part-ofspeech categories as well as other categories (family, sex, and death, to name but a few). One of LIWC's greatest strengths is that has been validated under different domains. For example, it has been used in opinion mining <ref type="bibr" target="#b5">[6]</ref>, in the analysis of suicide notes <ref type="bibr" target="#b6">[7]</ref>, in cyber-bullying detection <ref type="bibr" target="#b7">[8]</ref>, or satire identification <ref type="bibr" target="#b8">[9]</ref>. Although LIWC was originally conceived for the English language, it has a translated version available for Spanish. However, in contrast with LIWC, our proposal handles specific phenomena of the Spanish language such as grammatical gender, as well as drawing a fine-grained distinction among PoS categories, such as adjectives, adverbs, verbs, and suffixes, among others. A further strength of UMUTextStats over LIWC is that UMUTextStats allows for complex regular expressions that are helpful in order to capture complex features such as discourse markers, by means of which different arguments connected in a text can be captured. The current version of UMUTextStats captures a total of 311 different linguistic features organised in the following categories:</p><p>• Grammatical features (GRA). The features within this category are organised according to the Part-of-Speech (PoS) taxonomy. It includes verbs, adjectives, adverbs, determiners, pronouns, and conjunctions, to name but a few. Moreover, this category is organised into a more fine-grained distinction than LIWC. This decision was made because Spanish, as a highly inflected language, makes use of gender and number matching rules which may indicate when users are reporting facts, desires or hypothetical events.</p><p>• Morphological features (MOR). The features within this category extract information from components of the words including prefixes and suffixes (distinguishing between different types, namely denominal morphemes, deverbal morphemes, deadjectival morphemes, and evaluative suffixation by means of diminutives, augmentatives, and pejoratives). This category also contains features to match grammatical number (singular or plural).</p><p>• Spelling and stylistic mistakes (ERR). In this category, we distinguish between stylistic patterns, which denote the usage of colloquial language in order to detect atypical patterns, and linguistic errors, which can either denote the low cultural level of the writers, that is to say, errors in competence, or that they have failed to properly revise their writings before publishing them, that is, errors in performance.</p><p>• Figurative language (FIG). The features within this category are related to the usage of idioms, understatements, hyperboles or any other rhetorical device whose intention is to deviate the meaning of an utterance from its literal meaning <ref type="bibr" target="#b9">[10]</ref>.</p><p>• Linguistic processes (LPR). This category contains stylometric features such as number of words, syllables, and sentences. It also distinguishes between different types of sentences, such as interrogatory, exclamatory or literal quotation.</p><p>• Symbols (SYM). The features within this category aims to capture sentence dividers, such as spaces, colons, semicolons among other general-purpose symbols.</p><p>• Entity and topic extraction (ENT). These contain a list of general topics, including animals, food, jobs, clothes or body-parts, in order to determine the general topics of a text. In this category, we include the usage of inclusive language, analytic thinking, achievements and failures, and risk perception.</p><p>• Sentiments (SEN). The features within this category aims to determine general words and expressions related to positive and negative feelings.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Models based on word-embeddings</head><p>The state of the art regarding SA makes use of word-embeddings and deep-learning methods. Word-embeddings, compared with traditional methods such as those models based on n-grams and one-hot-vectors, have two significant advantages. On the one hand, they represent words as dense vectors rather than sparse vectors. The main idea beyond this approach is that dense vectors allow clustering words with similar meanings. On the other, word-embeddings can be initialised by applying unsupervised techniques from general purpose trained sources, such as social networks, free encyclopedias or news sites, instead of being initialised with random values. Moreover, the idea of word-embeddings could be extended from words to whole texts, in order to represent sentences of paragraphs as dense vectors <ref type="bibr" target="#b10">[11]</ref>. In this sense, sentence-embeddings are calculated by averaging the embeddings of the words that compose it.</p><p>Our participation in the TASS-2020 workshop involved three runs. The first run (LF + WE) consisted in linguistic features trained with a Multilayer Perceptron in combination with wordembeddings trained with a Convolutional Neural Network (CNN), the second run (LF) entailed the linguistic features trained with Support Vector Machines, and the third run (LF + SE) involved combining the linguistic features with sentence embeddings.</p><p>For the first run we used the functional API of Keras <ref type="bibr" target="#b11">[12]</ref> to create a classifier that combines the inputs from a CNN <ref type="bibr" target="#b12">[13]</ref>, composed by a Embedding Layer with Spanish pre-trained wordembeddings from fastText <ref type="bibr" target="#b13">[14]</ref> and a multilayer perceptron for training the linguistic features. The outputs for both deep-learning models where concatenated and combined with two more deep-learning layers and the the output layer for the final prediction. The architecture diagram of the first run is shown in Figure <ref type="figure" target="#fig_0">1</ref>. We proposed the usage of CNN in order to exploit the spatial dimension of word-embeddings because these kinds of networks are able to find common patterns of words regardless of their position in the text, which can help solve some NLP problems such as polysemy. In this run, the training dataset was used for training and the development dataset for evaluating our proposal.</p><p>For the second and the third runs, we used the Weka platform <ref type="bibr" target="#b14">[15]</ref> to evaluate the reliability of using linguistic features separately (LF) and in combination with sentence-embeddings (LF + SE). Both runs were trained with Sequential Minimal Optimisation (SMO), a machine-learning classifier which is based on Support Vector Machines (SVMs). Specifically, we set this SVM to use a polynomial kernel in order to learn from non-lineal models. These runs were trained with the combination of the training and development datasets.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Results</head><p>Our team participated in all the tasks proposed in the TASS-2020 workshop. All runs were evaluated using the macro-averaged versions of F-measure (F1), Precision (P), and Recall (R). First, Table <ref type="table" target="#tab_1">2</ref> contains the results of our runs for Subtask 1.1 regarding monolingual classification. For the first run, with the combination of CNN and linguistic features, the European Spanish dataset (es) achieved the best result with a macro F1-measure of 0.503311. As precision was higher than recall, we can assume that our proposal fails regarding class detection, but when this detection is successful, the results are quite reliable. However, the rest of the Spanish varieties achieved significantly worse results between a macro F1-measure of 0.322492 for the Costa Rica dataset and 0.391288 for the Uruguay dataset. We can also observe that precision and recall were more similar between them. However, for runs 2 and 3, where the training was performed by combining the training and development datasets, the results varied and both runs achieved their best result with the Uruguay dataset (0.49832 for run 2 and 0.520168 for run 3). However, with the European Spanish dataset the results are lower than the ones achieved for the first run, owing to the lack of the development dataset during the training stage.</p><p>When our results are compared with the rest of the participants, it will be observed that they achieved more stable results and their macro-F1 measure is much more similar across the different datasets. For example, the best result, achieved by daniel.palomino.paucar, obtained a macro-F1 measure of 0.64667, which outperforms the results from our proposal.</p><p>Next, the results achieved by our proposal for Subtask 1.2 regarding multilingual classification are shown in Table <ref type="table" target="#tab_2">3</ref>. In this subtask, we merged the datasets provided for Subtask 1.1 in order to create a multilingual classifier. It will be observed in the results that the runs for this subtask were similar to the ones achieved in Subtask 1.1, achieving our best macro F1-measure with 0.357876 for run 1, but obtaining similar precision and recall rates. The best result achieved in this subtask among all the participants was 0.497966. However, as the number of participants for this subtask was low, this result must be taken with caution. It is worth noting that the rules of the tasks allowed participants to use external corpora or linguistic resources. However, our participation was limited to the datasets provided. In this sense, we could not provide a fair comparison with the rest of the participants until we analyse their approaches in detail.</p><p>Compared with purely statistical models such as the ones based on word-embeddings, linguistic features provide interpretability. It is possible, therefore, to obtain the most discriminatory linguistic features by calculating Information Gain (IG), which is a metric used by ensemble methods such as decision trees in order to determine when a new branch must be created. Figure <ref type="figure" target="#fig_1">2</ref> shows the 20 best features with major information for the combination of the training datasets of the different datasets provided for Subtask 1.1. We can observe that Sentiments (SEN) is the linguistic category which provides the largest number of linguistic features including positive, negative, anger, offensive and sad. Out of these features, positive is, by far, the most discriminatory feature. Grammatical features (GRA) is another linguistic feature that provides several features, such as adverbs-negation and adjectives-qualifying. It is worth noting that these linguistic features are hard to obtain by applying models based on word-embeddings like those linguistic features regarding stylistic patterns (misspellings, linguistic errors, and swear), as well as other features used to add emphasis such as the number of exclamatory sentences.  Finally, the results from Task 2, emotion detection, are shown in Table <ref type="table" target="#tab_3">4</ref>. As can be observed, the usage of linguistic features and sentence embeddings (run 3) achieved the best result over the rest of our runs. It worth noting that run 3 only achieves slightly better results than the run 2 (LF), but with lower precision and higher recall. Our best macro F1-measure is 0.378694, which is reasonably successful, in view of the complexity of the task although the best score was achieved by jogonba2elirf with a macro F1-measure of 0.446582.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusions and further work</head><p>In this paper, the participation of the UMUTeam in the TASS'2020 Workshop on Sentiment Classification applying linguistic features has been described. For Subtask 1.1, our proposal achieved decent results only for one dataset: European Spanish with CNN and linguistic features, and for the Uruguay dataset with SMO and linguistic features with average word-embeddings. For Subtask 1.2, which consisted in the combination of the datasets from different Spanish varieties, we achieved more stable results irrespective of the model applied. However, the low number of participants hindered the achievement of better insights. Finally, for Task 2, which consisted in emotion detection, we achieved reasonably good results with a macro F1 of 0.372774 over six different classes.</p><p>As regards RQ1 and RQ2, we have observed that our proposal achieves good results when comparing the usage of LF separately with the rest of our runs, but the results are far from being the best results. In this sense, we need to perform a more detailed analysis of the rest of the participants in order to identify the weakness of our proposal. Regarding RQ3, we have observed that our proposal does not provide stable results for Subtask 1.1 when the different dialects are considered separately, but the results were more stable when training and testing contained tweets from the different dialects as the same time. This fact suggests that some of the linguistic features between each dialect and cultural background are complementary, but not enough to perform a reliable sentiment classification.</p><p>We are well satisfied with our participation for the first time in a TASS workshop and our competition in challenging NLP tasks. We are, however, aware of the limitations of our proposal and the long way to go. To our mind, the main drawback is that we focused only on European Spanish during the design of the linguistic features. We will, therefore, focus on improvements for other varieties enabling the adaptation of the system. Regarding the technological aspect, we will include the hyper-parameter tuning in our pipeline, in order to choose the optimal hyper-parameters for a learning algorithm. We will also try to incorporate other pre-trained and contextualised word-embeddings such as ELMo <ref type="bibr" target="#b15">[16]</ref> and BERT <ref type="bibr" target="#b16">[17]</ref>.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Model architecture of the first run, consisting in the combination of a Convolutional Neural Work for word-embeddings and a Multilayer perceptron for the linguistic features</figDesc><graphic coords="6,141.38,84.19,312.52,303.52" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Information gain of the 20 best linguistic features for the Task 1.2</figDesc><graphic coords="8,141.38,106.52,312.52,238.36" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Corpus size distribution for each dataset of the TASS'2020 corpus</figDesc><table><row><cell>Model</cell><cell cols="4">Training Development Testing Total</cell></row><row><cell>Subtask 1.1 (es)</cell><cell>1127</cell><cell>582</cell><cell>1707</cell><cell>3416</cell></row><row><cell>Subtask 1.1 (pe)</cell><cell>967</cell><cell>499</cell><cell>1465</cell><cell>2931</cell></row><row><cell>Subtask 1.1 (cr)</cell><cell>778</cell><cell>391</cell><cell>1167</cell><cell>2336</cell></row><row><cell>Subtask 1.1 (uy)</cell><cell>944</cell><cell>487</cell><cell>1429</cell><cell>2860</cell></row><row><cell>Subtask 1.1 (mx)</cell><cell>991</cell><cell>511</cell><cell>1501</cell><cell>3003</cell></row><row><cell>Subtask 1.2</cell><cell>0</cell><cell>0</cell><cell>1501</cell><cell>1501</cell></row><row><cell>Subtask 2</cell><cell>5886</cell><cell>857</cell><cell>1666</cell><cell>8409</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>Results of the UMUTeam in Task 1.1 Monolingual classification</figDesc><table><row><cell>Model</cell><cell>Dataset</cell><cell>F1</cell><cell>P</cell><cell>R</cell></row><row><cell cols="2">CNN (LF + WE) es</cell><cell cols="3">0.503311 0.561294 0.456186</cell></row><row><cell cols="2">CNN (LF + WE) pe</cell><cell>0.379346</cell><cell>0.399611</cell><cell>0.361037</cell></row><row><cell cols="2">CNN (LF + WE) cr</cell><cell>0.322492</cell><cell>0.321570</cell><cell>0.323420</cell></row><row><cell cols="2">CNN (LF + WE) uy</cell><cell>0.391288</cell><cell>0.393571</cell><cell>0.389031</cell></row><row><cell cols="2">CNN (LF + WE) mx</cell><cell>0.373712</cell><cell>0.377036</cell><cell>0.370446</cell></row><row><cell>SMO (LF)</cell><cell>es</cell><cell>0.373223</cell><cell>0.373757</cell><cell>0.372691</cell></row><row><cell>SMO (LF)</cell><cell>pe</cell><cell>0.379505</cell><cell>0.393181</cell><cell>0.366748</cell></row><row><cell>SMO (LF)</cell><cell>cr</cell><cell>0.333147</cell><cell>0.333461</cell><cell>0.332834</cell></row><row><cell>SMO (LF)</cell><cell>uy</cell><cell cols="3">0.498352 0.498630 0.498075</cell></row><row><cell>SMO (LF)</cell><cell>mx</cell><cell>0.373502</cell><cell>0.373762</cell><cell>0.373241</cell></row><row><cell>SMO (LF + SE)</cell><cell>es</cell><cell>0.378384</cell><cell>0.378490</cell><cell>0.378278</cell></row><row><cell>SMO (LF + SE)</cell><cell>pe</cell><cell>0.389992</cell><cell>0.396927</cell><cell>0.383295</cell></row><row><cell>SMO (LF + SE)</cell><cell>cr</cell><cell>0.349812</cell><cell>0.349505</cell><cell>0.350119</cell></row><row><cell>SMO (LF + SE)</cell><cell>uy</cell><cell cols="3">0.520168 0.518072 0.522281</cell></row><row><cell>SMO (LF + SE)</cell><cell>mx</cell><cell>0.396592</cell><cell>0.397157</cell><cell>0.396030</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc>Results of the UMUTeam in Task 1.2: Multilingual classification</figDesc><table><row><cell>Model</cell><cell>F1</cell><cell>P</cell><cell>R</cell></row><row><cell cols="2">CNN (LF + WE) 0.336824</cell><cell>0.337451</cell><cell>0.336198</cell></row><row><cell>SMO (LF)</cell><cell cols="3">0.357876 0.358662 0.357093</cell></row><row><cell>SMO (LF + SE)</cell><cell>0.334466</cell><cell>0.335797</cell><cell>0.333146</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4</head><label>4</label><figDesc>Results of the UMUTeam in Task 2: Emotion detection</figDesc><table><row><cell>Model</cell><cell>F1</cell><cell>P</cell><cell>R</cell></row><row><cell cols="2">CNN (LF + WE) 0.242799</cell><cell>0.243873</cell><cell>0.241734</cell></row><row><cell>SMO (LF)</cell><cell cols="3">0.372774 0.490520 0.300614</cell></row><row><cell>SMO (LF + SE)</cell><cell cols="3">0.378694 0.419558 0.345084</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">http://aspell.net/</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work has been supported by the Spanish National Research Agency (AEI) and the European Regional Development Fund (FEDER/ERDF) through projects KBS4FIA (TIN2016-76323-R) and LaTe4PSP (PID2019-107652RB-I00). In addition, José Antonio García-Díaz has been supported by Banco Santander and University of Murcia through the Doctorado industrial programme.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Lie catching and microexpressions</title>
		<author>
			<persName><forename type="first">P</forename><surname>Ekman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The philosophy of deception</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page">5</biblScope>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Overview of tass 2020: Introduction emotion detection</title>
		<author>
			<persName><forename type="first">M</forename><surname>García-Vega</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">C</forename><surname>Díaz-Galiano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>García-Cumbreras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Montejo Ráez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Jiménez Zafra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Martínez-Cámara</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">A</forename><surname>Murillo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Murillo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Chiruzzo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Moctezuma</surname></persName>
		</author>
		<author>
			<persName><surname>Sobrevilla</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Iberian Languages Evaluation Forum</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<meeting>the Iberian Languages Evaluation Forum<address><addrLine>IberLEF; Málaga, Spain</addrLine></address></meeting>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2020">2020. 2020</date>
			<biblScope unit="volume">2664</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">EmoEvent: A multilingual emotion corpus based on different events</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">M</forename><surname>Del Arco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Strapparava</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">A</forename><surname>Urena Lopez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Martin</surname></persName>
		</author>
		<ptr target="https://www.aclweb.org/anthology/2020.lrec-1.186" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of The 12th Language Resources and Evaluation Conference, European Language Resources Association</title>
				<meeting>The 12th Language Resources and Evaluation Conference, European Language Resources Association<address><addrLine>Marseille, France</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="1492" to="1498" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Ontology-driven aspect-based sentiment analysis classification: An infodemiological case study regarding infectious diseases in latin america</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>García-Díaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Cánovas-García</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Valencia-García</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.future.2020.06.019</idno>
		<idno>doi:</idno>
		<ptr target="https://doi.org/10.1016/j.future.2020.06.019" />
	</analytic>
	<monogr>
		<title level="j">Future Generation Computer Systems</title>
		<imprint>
			<biblScope unit="volume">112</biblScope>
			<biblScope unit="page" from="614" to="657" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">The psychological meaning of words: Liwc and computerized text analysis methods</title>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">R</forename><surname>Tausczik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">W</forename><surname>Pennebaker</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of language and social psychology</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="page" from="24" to="54" />
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">A study on LIWC categories for opinion mining in spanish reviews</title>
		<author>
			<persName><forename type="first">M</forename><surname>Del</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pilar</forename><surname>Salas-Zárate</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>López-López</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Valencia-García</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Aussenac-Gilles</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Á</forename><surname>Almela</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Alor-Hernández</surname></persName>
		</author>
		<idno type="DOI">10.1177/0165551514547842</idno>
		<ptr target="https://doi.org/10.1177/0165551514547842.doi:10.1177/0165551514547842" />
	</analytic>
	<monogr>
		<title level="j">J. Inf. Sci</title>
		<imprint>
			<biblScope unit="volume">40</biblScope>
			<biblScope unit="page" from="749" to="760" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">A linguistic analysis of suicide-related twitter posts</title>
		<author>
			<persName><forename type="first">B</forename><surname>O'dea</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">E</forename><surname>Larsen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">J</forename><surname>Batterham</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">L</forename><surname>Calear</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Christensen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Crisis: The Journal of Crisis Intervention and Suicide Prevention</title>
		<imprint>
			<biblScope unit="volume">38</biblScope>
			<biblScope unit="page">319</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Toward multimodal cyberbullying detection</title>
		<author>
			<persName><forename type="first">V</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ghosh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Jose</surname></persName>
		</author>
		<idno type="DOI">10.1145/3027063.3053169</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2017 CHI Conference Extended Abstracts on Human Factors in Computing Systems</title>
				<meeting>the 2017 CHI Conference Extended Abstracts on Human Factors in Computing Systems</meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="volume">127655</biblScope>
			<biblScope unit="page" from="2090" to="2099" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Automatic detection of satire in twitter: A psycholinguisticbased approach</title>
		<author>
			<persName><forename type="first">M</forename><surname>Del</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pilar</forename><surname>Salas-Zárate</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Paredes-Valverde</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">Á</forename><surname>Rodríguez-García</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Valencia-García</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Alor-Hernández</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.knosys.2017.04.009</idno>
		<ptr target="https://doi.org/10.1016/j.knosys.2017.04.009.doi:10.1016/j.knosys.2017.04.009" />
	</analytic>
	<monogr>
		<title level="j">Knowl. Based Syst</title>
		<imprint>
			<biblScope unit="volume">128</biblScope>
			<biblScope unit="page" from="20" to="33" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Review of english literature on figurative language applied to social networks</title>
		<author>
			<persName><forename type="first">M</forename><surname>Del</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pilar</forename><surname>Salas-Zárate</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Alor-Hernández</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">L</forename><surname>Sánchez-Cervantes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Paredes-Valverde</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">L</forename><surname>García-Alcaraz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Valencia-García</surname></persName>
		</author>
		<idno type="DOI">10.1007/s10115-019-01425-3</idno>
		<ptr target="https://doi.org/10.1007/s10115-019-01425-3.doi:10.1007/s10115-019-01425-3" />
	</analytic>
	<monogr>
		<title level="j">Knowl. Inf. Syst</title>
		<imprint>
			<biblScope unit="volume">62</biblScope>
			<biblScope unit="page" from="2105" to="2137" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><surname>Cer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>-Y. Kong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Hua</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Limtiaco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">S</forename><surname>John</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Constant</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Guajardo-Cespedes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Yuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Tar</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1803.11175</idno>
		<title level="m">Universal sentence encoder</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">F</forename><surname>Chollet</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2015">2015. 2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">Convolutional neural networks for sentence classification</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Kim</surname></persName>
		</author>
		<idno>CoRR abs/1408.5882</idno>
		<ptr target="http://arxiv.org/abs/1408.5882.arXiv:1408.5882" />
		<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">Learning word vectors for 157 languages</title>
		<author>
			<persName><forename type="first">E</forename><surname>Grave</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Bojanowski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Joulin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Mikolov</surname></persName>
		</author>
		<idno>CoRR abs/1802.06893</idno>
		<ptr target="http://arxiv.org/abs/1802.06893.arXiv:1802.06893" />
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">The weka data mining software: an update</title>
		<author>
			<persName><forename type="first">M</forename><surname>Hall</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Frank</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Holmes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Pfahringer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Reutemann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">H</forename><surname>Witten</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM SIGKDD explorations newsletter</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="page" from="10" to="18" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">E</forename><surname>Peters</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Neumann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Iyyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gardner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Clark</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</author>
		<idno>CoRR abs/1802.05365</idno>
		<ptr target="http://arxiv.org/abs/1802.05365.arXiv:1802.05365" />
		<title level="m">Deep contextualized word representations</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-W</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1810.04805</idno>
		<title level="m">Bert: Pre-training of deep bidirectional transformers for language understanding</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
