<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">HAHA at EmoEvalEs 2021: Sentiment Analysis in Spanish Tweets with Cross-lingual Model</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Kun</forename><surname>Li</surname></persName>
							<email>2106967047@qq.com</email>
							<affiliation key="aff0">
								<orgName type="department">School of Information Science and Engineering</orgName>
								<orgName type="institution">Yunnan University</orgName>
								<address>
									<settlement>Yunnan</settlement>
									<country key="CN">P.R. China</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">HAHA at EmoEvalEs 2021: Sentiment Analysis in Spanish Tweets with Cross-lingual Model</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">4AAB86A5B4C49D032203612EC13B875F</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T00:23+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Sentiment Analysis</term>
					<term>Data Augmentation</term>
					<term>Machine Learning</term>
					<term>Cross-lingual Model</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This article describes HAHATeam participation in the Iber-LEF 2021 EmoEvalEs task: Emotion detection and Evaluation for Spanish. In this task, we use Masked Language Model-based data augmentation to enhance the data to increase the training data set and prevent overfitting. We improve the performance of the model through a series of data processing and data augmentation technologies. We use three different cross-language models, BERT, XLM, and XLM-RoBERTa for comparative experiments. Very competitive results have been achieved on both the development set and the test set. The best model achieved the third place in the development set and the fifth place in the test set.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Sentiment analysis (SA) is the process of analyzing, processing and categorizing subjective text. According to the emotional tendency of the text, the text can be classified. In addition to the traditionally recognized positive and negative emotions, there should also be anger, disgust, fear, joy, sadness, surprise and other emotion types. This article describes the emotion classification task (EmoEvalEs <ref type="bibr" target="#b10">[11]</ref>) that we participated in in IberLEF 2021 <ref type="bibr" target="#b6">[7]</ref>, which requires the classification of Spanish tweets into various categories based on emotions, such as anger, disgust, fear, joy, sadness, surprise, or others. Due to the lack of voice regulation and facial expressions, understanding the emotions expressed by users on social media is a difficult task.</p><p>The earliest work on SA largely relied on feature engineering <ref type="bibr" target="#b14">[15,</ref><ref type="bibr" target="#b2">3]</ref>, and subsequent neural network-based methods <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b16">17,</ref><ref type="bibr" target="#b13">14,</ref><ref type="bibr" target="#b15">16</ref>] achieved higher accuracy. Recently, Ma <ref type="bibr" target="#b5">[6]</ref> said: "Incorporate useful common sense knowledge into deep neural networks to further enhance the results of the model." Liu <ref type="bibr" target="#b4">[5]</ref> optimized the memory network and applied it to its model to better capture the language structure.</p><p>Recently, pre-trained language models, such as ELMo <ref type="bibr" target="#b9">[10]</ref>, OpenAIGPT <ref type="bibr" target="#b12">[13]</ref>, and BERT <ref type="bibr" target="#b1">[2]</ref>, have shown their effectiveness in mitigating feature engineering efforts. In particular, BERT has achieved excellent results in Question Answering (QA) and Natural Language Inference (NLI). However, the SA task that directly uses the pre-trained BERT model does not improve much.</p><p>In order to solve this task, we proposed the use of pretrained classic deep learning models to apply to downstream task. Furthermore, we not only tested the data on a single model, but also compared three language models. We use the three mainstream cross-language models: bert-multiling-cased (BERT), xlmmlm-100-1280 (XLM), xlm-roberta-base (XLM-RoBERTa). These three models all come from the transformers package, and the hugging face provided us with great help. The system will sort according to accuracy and the weighted-averaged versions of Precision, Recall, and F1.</p><p>The rest of the article is structured as follows. Section2 proposes the task to be solved and introduces the corpus of the task. It analyzes the data set and proposes some tricks. In the Section3, we introduce the methods and techniques we used. The first is to use some conventional operations to process the data set, the second is to introduce the data augmentation method we use, and the third is to introduce the language model we use. Section 4 presents the results and discuss the performance of the system. In Section 5, we give the conclusions and future work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Task and Corpus description</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Task description</head><p>The organizer only proposed one task: EmoEvalEs <ref type="bibr" target="#b10">[11]</ref>: Emotion detection and Evaluation for Spanish. Understanding the emotions expressed by users on social media is a hard task due to the absence of voice modulations and facial expressions. Our shared task "Emotion detection and Evaluation for Spanish" has been designed to encourage research in this area. The task consists of classifying the emotion expressed in a tweet as one of the following emotion classes: Anger, Disgust, Fear, Joy, Sadness, Surprise or Others.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Corpus description</head><p>The datasets <ref type="bibr" target="#b11">[12]</ref> is based on events related to different fields that occurred in April 2019: entertainment, disasters, politics, global memorials, and global strikes. The corpus contains approximately 8.2k tweets for this task. Each tweet contains 5 features: id, event, tweet, offensive and emotion. The most important feature is tweet. At the same time, event and offensive may also have a positive impact on the final result. We will verify our conjecture in the next part of the experiment. For this task, the corpus is released as three data sets, namely training, development and testing. Table <ref type="table" target="#tab_0">1</ref> shows the distribution of training set and development set. 3 Method and technology</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Data Preprocessing</head><p>To achieve good results, data preprocessing is essential. The first step in our model is to process these tweets. For how to preprocess the data, we refer to IberLEF 2019 <ref type="bibr" target="#b8">[9]</ref>, which gives us some ideas and and more in-depth analysis. We found that in addition to tweets, there are two other features that may affect that final result, so we process the datasets as follows:</p><p>-Combining tweet, event, and offensive together as a new text. The advantage of this is to make better use of the information given by the organizer. -Removing URLs.</p><p>-Contracting white-space characters, such as space, tabs or punctuation.</p><p>-Removing stop word in Spanish.</p><p>We ignored spelling errors in text tweets. After data processing, the sentence length in the tweet is distributed within 60. At the same time, we also keep the version that not remove the stop word in Spanish, so that we can judge whether removing the stop word in Spanish is good for our model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Data Augmentation</head><p>Data augmentation technology is widely used in computer vision, but it is rarely effectively used in Natural Language Processing (NLP). The essential reason is that some data augmentation methods in the image will not change the meaning of the image itself while augmenting the data. In this article, we try to use a data augmentation method: Masked Language Model-based data augmentation. Masked Language Modeling (MLM) can not only predict words, but also can be used for text data augmentation, and other methods, such as based on synonym substitution and word embedding. Compared with the replacement of, the text generated by data augmentation based on the Masked Language Model is more grammatically fluent, because the model considers the context information when making predictions. We can easily use HuggingFace's transformers library. One place to pay attention to in this method is how to determine which part of the Mask text is to be used. Do data augmentation for each processed tweet, and determine the number of masks according to the length of each sentence, details as follows: "score" indicates the degree of similarity between the word and [Mask], and "sequence" indicates the sequence after data augmentation. The sequence with the highest score is selected as the sentence augmented by our data.</p><formula xml:id="formula_0">-Fixed 1 [Mask],</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Model</head><p>For SA Task, we combine some classic deep learning models. In the deep learning model, we mainly use the most popular cross-lingual models. These cross-lingual models can be better used to process Spanish text. These models include two parts: tokenizer and SequenceClassification. Among them, the tokenizer divides the input into words and processes it into the input required by the model. SequenceClassification is applied to specific downstream tasks.</p><p>tokenizer: First of all, the tokenizer was splitting strings in sub-word token strings, converting tokens strings to ids and back, and encoding or decoding. Adding new tokens to the vocabulary in a way that is independent of the underlying structure (Byte Pair Encoding). Managig special tokens (like mask, begining-of-sentence, etc). The model needs language embedding during inference. We will use the tokenizer module to process the input before entering the model. The input required for each cross-language model is different, and the tokenizer module will generate the corresponding embedding according to the specific model. The tweet first undergoes data preprocessing, and then uses the tokenizer to vectorize the input. According to different models, Token embeddings, Segment Embeddings, and Position Embeddings are obtained.</p><p>-SequenceClassification: The model with a sequence classification/regression heads on top(a linear layer on top of the pooled output). We add a dropout layer and a linear layer to the last layer. The input feature of the linear layer we added is 768, the output feature is 384, and the p value of the dropout layer is set to 0.1. The input feature of the last layer is 384, and the output feature is 7, to satisfy our 7 classification task. The model does not require language embedding at inference time. They can identify the language used in the context and infer accordingly.</p><p>XLM-RoBERTa <ref type="bibr" target="#b0">[1]</ref>: The XLM-RoBERTa model was proposed in Unsupervised Cross-lingual Representation Learning. It is a large multi-lingual language model, trained on 2.5TB of filtered Common Crawl data. XLM-RoBERTa is a multilingual model trained on 100 different languages. Unlike some XLM multilingual models, it does not require lang tensors to understand which language is used, and should be able to determine the correct language from the inputids. XLM-RoBERTa was evaluated on classification tasks and they showed very good performance. During fine-tuning, multi-language annotation data is used based on the ability of the multi-language model to improve the performance of downstream tasks. This allows our model to obtain state-of-the-art results in cross-language benchmark tests, while surpassing the performance of the singlelanguage BERT model in each language. The main body of the XLM-RoBERTa model is Transformer, and the training target is the multilingual MLM objective, which is basically the same as XLM. We sample the text from each language, and then predict the tokens that are masked. We use model Sentence Piece with unigram language model to perform Sentence Piece on the original text data.</p><p>XLM <ref type="bibr" target="#b3">[4]</ref>: Although the BERT model can be pre-trained on hundreds of languages, the information between languages is not interoperable, and there is no shared knowledge between different language models. Facebook's XLM model overcomes the problem of non-interoperability of information, puts different languages together and uses new training targets for training, so that the model can master more cross-language information. A significant advantage of this cross-language model is that, for subsequent tasks after pre-training (such as text classification or translation tasks), languages with relatively rare training corpus can use the information learned on other corpora. We propose to use xlm-mlm-100-1280 as part of our experiment.</p><p>BERT: The bert-base-multilingual-cased model is based on BERT, which has obtained state-of-the-art performance on most NLP task <ref type="bibr" target="#b1">[2]</ref>. The pre-trained BERT models are trained on a large corpus (Wikipedia). Trained on cased text in the top 104 languages with the largest Wikipedias. So bert-base-multilingualcased model is better for Spanish text in Sentiment Analysis (SA). Because bertbase-multilingual-cased model can splits tokens in its vocabulary into sub-tokens, while will affect the result of the classification task. The bert-base-multilingualcased model also fixes normalization issues in many languages, so it is recommended in languages with non-Latin alphabets (and is often better for most languages with Latin alphabets).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Experiments</head><p>We have fully experimented with the above three model. The process and result of the experiment will be described in detail below. We first experiment with different data augmentation methods to obtain better data augmentation methods. Then we use the best data augmentation method on the three crosslanguage models and compare them on the development set. Finally, we use the best model to experiment on the test set and get the final result.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Experimental Setup</head><p>During the development set experiment, we divided the training set into a new training set and a development set, where the division ratio was 0.9 for the new training set, and the development set was used as the test set.</p><p>We will do the same preprocessing before sending the data to each model to ensure the comparability of the experiment, that is the ablation experiment. Use ablation experiments to evaluate the quality of our different data processing methods. And here some hyper parameter sett of the model; the max length of a single sentence in each batch is adjusted according to the preprocessed results, the max length does not exceed 60; and add "[CLS]" and "[SEP]" to each input; and padding for sentences with insufficient length; We use a batch size of 32, a base learning rate of 1e-5; the activation function is AdamW; and the number of fine-tuning is set to 3. For every 10 batch-size training set data processed, model.eval will be executed to process the development set data. And save the model and print the accuracy rate.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Learning Rate Warm-up</head><p>Use learning rate warm-up. During warm-up, the learning rate linearly increases from 0 to the initial learning rate of the optimizer. After the warm-up phase, a schedule is created to linearly reduce the learning rate from the initial learning rate in the optimizer to 0. The number of steps for the warm up phase is set to 0, that is, the learning rate is warmed up from the beginning. The total number of training steps are set to N.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>N = len(train dataloader) × num epochs</head><p>(1)</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Data Augmentation Experiments</head><p>We conduct data augmentation experiments on the XLM-RoBERTa model to get the best data augmentation method, and use the data augmentation method in all subsequent models. From Table <ref type="table" target="#tab_2">2</ref>, we can see that when using 15% of </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4">Finally Experiments</head><p>Next, we conduct ablation experiments on three cross-language models for data augmentation. From Table <ref type="table" target="#tab_3">3</ref>, we can see that the four evaluation indicators of the BERT model and the XLM model during the data augmentation operation are less than those without the data augmentation operation. In the XLM-RoBERTa model, data augmentation can improve the performance of the model to a certain extent, except for Precision weighted scores. Furthermore, whether data augmentation is performed or not, the performance of the XLM-RoBERTa model is better than the other two models. The biggest advantage of XLM-RoBERTa is the significant increase in the amount of training data. And the effect is especially obvious in languages with fewer resources. Finally, we use the best performing XLM-RoBERTa model on the test data set, with data augmentation and no data augmentation at the same time. From Table <ref type="table" target="#tab_4">4</ref>, we can see that data augmentation will effectively improve the performance of the model in test data set, especially in Accuracy, Recall weighted and F1 weighted scores. But it will also reduce the Precision weighted score of the model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.5">Result Analysis</head><p>We use three popular cross-language models BERT, XLM, and XLM-RoBERTa for comparative experiments. It can be seen from Figure <ref type="figure">3</ref> that the XLM-RoBERTa model surpasses BERT and XLM in all evaluation indicators. Because the XLM-RoBERTa model increases the number of languages and the number of training data sets, it uses more than 2TB of pre-processed CommonCrawl data sets. And during the fine-tuning of the XLM-RoBERTa model, based on the ability of the multi-language model to use multi-language annotation data to improve the performance of downstream tasks. Beyond this, although we use data augmentation technology to only double the amount of data, the performance improvement is still very obvious. Accuracy and Recall weighted have increased by 0.6%, and F1 weighted has increased by 1.2%. Moreover, the data processing method and Learning Rate Warm-up technology we use also play a certain role in the improvement of model performance and the speed of model inference. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Conclusions and further work</head><p>This article introduces the EmoEvalEs task: Emotion detection and Evaluation for Spanish in the IberLEF 2021. We use three cross-language models BERT, XLM and XLM-RoBERTa. Using the cross-language pre-training model can handle Spanish tweets well, and the pre-training model only needs to modify the final output layer to be well applied to downstream tasks, which greatly reduces the training time. We improve the performance of the model through a series of data processing and data augmentation technologies. As can be seen from Table <ref type="table" target="#tab_5">5</ref>, The results and ranking of our model on the official test sets (Ranked by Accuracy). Our final model got an Accuracy of 0.692, a Precision weighted score of 0.680, a Recall weighted score of 0.692, and an F1 weighted score of 0.664 on the official test set, ranking 5th, 6th, 5th, and 8th among all participating teams. We are very satisfied with our first participation in IberLEF 2021 and our performance in the EmoEvalEs task, but some work still needs improvement. The future work is to explore more advanced technology, use better models, and achieve better results in the next competition. For example, use better data augmentation technology to increase the amount of data set, select the best hyperparameter adjustment technology, and use other pre-training models such as Longformer, ELECTRA, ALBERT or even T5.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 .</head><label>1</label><figDesc>Distribution of tweets in the training and development set of the task.</figDesc><table><row><cell cols="7">Joy Sadness Anger Surprise Disgust Fear Others Total</cell></row><row><cell cols="2">training 1408 797</cell><cell>674</cell><cell>273</cell><cell>127</cell><cell cols="2">74 3214 5723</cell></row><row><cell>development 181</cell><cell>104</cell><cell>85</cell><cell>35</cell><cell>16</cell><cell>9</cell><cell>414 844</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head></head><label></label><figDesc>random position, repeatable.-Fixed 2 [Mask], random position, repeatable.-Fixed 3 [Mask], random position, repeatable.-15% of the length of the sentence [Mask], the position is random, not repeatable. -20% of the length of the sentence [Mask], the position is random, not repeatable.</figDesc><table><row><cell>Next is a concrete example: 'This is [Mask] cool'</cell></row><row><cell>'score': 0.515411913394928, 'sequence': 'This is [pretty] cool'</cell></row><row><cell>'score': 0.1166248694062233, 'sequence': 'This is [really] cool'</cell></row><row><cell>'score': 0.07387523353099823, 'sequence': 'This is [super] cool'</cell></row><row><cell>'score': 0.04272908344864845, 'sequence': 'This is [kinda] cool'</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2 .</head><label>2</label><figDesc>The results of different data augmentation strategies on the development data set on the XLM-RoBERTa model. RoBERTa is the best among the three evaluation indicators Accuracy, Recall weighted and F1 weighted score. But when the XLM-RoBERTa model uses 20% of the sentence length [Mask], the precision of the model is the best. The choice of pre-training model has a great impact on the final performance, even though they are all based on transformers.</figDesc><table><row><cell cols="5">XLM-RoBERTa Accuracy Precision weighted Recall weighted F1 weighted</cell></row><row><cell>Fixed 1</cell><cell>0.697</cell><cell>0.671</cell><cell>0.697</cell><cell>0.680</cell></row><row><cell>Fixed 2</cell><cell>0.698</cell><cell>0.670</cell><cell>0.698</cell><cell>0.678</cell></row><row><cell>Fixed 3</cell><cell>0.687</cell><cell>0.657</cell><cell>0.687</cell><cell>0.666</cell></row><row><cell>15% random</cell><cell>0.710</cell><cell>0.676</cell><cell>0.710</cell><cell>0.685</cell></row><row><cell>20% random</cell><cell>0.703</cell><cell>0.677</cell><cell>0.703</cell><cell>0.681</cell></row><row><cell cols="3">the length of the sentence [Mask], XLM-</cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 3 .</head><label>3</label><figDesc>The results of different data augmentation strategies on the development data set on the cross-language model.</figDesc><table><row><cell>Model</cell><cell cols="5">Data Augmentation Accuracy Precision weighted Recall weighted F1 weighted</cell></row><row><cell>BERT</cell><cell>Yes No</cell><cell>0.626 0.666</cell><cell>0.606 0.618</cell><cell>0.626 0.666</cell><cell>0.614 0.637</cell></row><row><cell>XLM</cell><cell>Yes No</cell><cell>0.659 0.684</cell><cell>0.654 0.673</cell><cell>0.659 0.685</cell><cell>0.656 0.669</cell></row><row><cell>XLM-RoBERTa</cell><cell>Yes No</cell><cell>0.710 0.704</cell><cell>0.676 0.702</cell><cell>0.710 0.704</cell><cell>0.685 0.677</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 4 .</head><label>4</label><figDesc>The results of the XLM-RoBERTa model on the test set.</figDesc><table><row><cell>Model</cell><cell cols="5">Data Augmentation Accuracy Precision weighted Recall weighted F1 weighted</cell></row><row><cell>XLM-RoBerta</cell><cell>Yes No</cell><cell>0.698 0.692</cell><cell>0.647 0.680</cell><cell>0.698 0.692</cell><cell>0.670 0.664</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 5 .</head><label>5</label><figDesc>The results and ranking of our model on the official test sets (Ranked by Accuracy).</figDesc><table><row><cell cols="6">Team Accuracy Precision weighted Recall weighted F1 weighted rank</cell></row><row><cell cols="2">GSI-UPM 0.728</cell><cell>0.709</cell><cell>0.728</cell><cell>0.717</cell><cell>1</cell></row><row><cell>HAHA</cell><cell>0.698</cell><cell>0.647</cell><cell>0.698</cell><cell>0.670</cell><cell>5</cell></row><row><cell>qu</cell><cell>0.450</cell><cell>0.619</cell><cell>0.450</cell><cell>0.447</cell><cell>15</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>First of all, we would like to thank the organizers for the opportunity and organization, as well as the help of teachers and seniors. Finally, we would like to thank the school for supporting my research and the patient work of future reviewers.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">Unsupervised cross-lingual representation learning at scale</title>
		<author>
			<persName><forename type="first">A</forename><surname>Conneau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Khandelwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Chaudhary</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Wenzek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Guzmán</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Grave</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ott</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Stoyanov</surname></persName>
		</author>
		<idno>CoRR abs/1911.02116</idno>
		<ptr target="http://arxiv.org/abs/1911.02116" />
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">BERT: pre-training of deep bidirectional transformers for language understanding</title>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
		<idno>CoRR abs/1810.04805</idno>
		<ptr target="http://arxiv.org/abs/1810.04805" />
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">NRC-Canada-2014: Detecting aspects and sentiment in customer reviews</title>
		<author>
			<persName><forename type="first">S</forename><surname>Kiritchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Cherry</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mohammad</surname></persName>
		</author>
		<idno type="DOI">10.3115/v1/S14-2076</idno>
		<ptr target="https://www.aclweb.org/anthology/S14-2076" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 8th International Workshop on Semantic Evaluation</title>
				<meeting>the 8th International Workshop on Semantic Evaluation<address><addrLine>SemEval; Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2014-08">2014. Aug 2014</date>
			<biblScope unit="page" from="437" to="442" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Cross-lingual language model pretraining</title>
		<author>
			<persName><forename type="first">G</forename><surname>Lample</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Conneau</surname></persName>
		</author>
		<idno>CoRR abs/1901.07291</idno>
		<ptr target="http://arxiv.org/abs/1901.07291" />
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Recurrent entity networks with delayed memory update for targeted aspect-based sentiment analysis</title>
		<author>
			<persName><forename type="first">F</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Cohn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Baldwin</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/N18-2045</idno>
		<ptr target="https://www.aclweb.org/anthology/N18-2045" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</title>
				<meeting>the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies<address><addrLine>New Orleans, Louisiana</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2018-06">Jun 2018</date>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="278" to="283" />
		</imprint>
	</monogr>
	<note>Short Papers</note>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Peng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Cambria</surname></persName>
		</author>
		<ptr target="https://www.aaai.org/ocs/index.php/AAAI/AAAI18/paper/view/16541" />
		<title level="m">Targeted aspect-based sentiment analysis via embedding commonsense knowledge into an attentive lstm</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Montes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Aragón</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Agerri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">Á</forename><surname>Álvarez-Carmona</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Álvarez Mellado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Carrillo-De Albornoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Chiruzzo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Freitas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Gómez Adorno</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Gutiérrez</surname></persName>
		</author>
		<title level="m">Proceedings of the Iberian Languages Evaluation Forum</title>
				<editor>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Jiménez-Zafra</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Lima</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><forename type="middle">M</forename><surname>Plaza-De Arco</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Taulé</surname></persName>
		</editor>
		<meeting>the Iberian Languages Evaluation Forum<address><addrLine>IberLEF</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021. 2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">PhraseRNN: Phrase recursive neural network for aspect-based sentiment analysis</title>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">H</forename><surname>Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Shirai</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/D15-1298</idno>
		<ptr target="https://www.aclweb.org/anthology/D15-1298" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing</title>
				<meeting>the 2015 Conference on Empirical Methods in Natural Language Processing<address><addrLine>Lisbon, Portugal</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2015-09">Sep 2015</date>
			<biblScope unit="page" from="2509" to="2514" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">Retuyt-inco at tass 2019: Sentiment analysis in spanish tweets</title>
		<author>
			<persName><forename type="first">M</forename><surname>Pastorini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Pereira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Zeballos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Chiruzzo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rosá</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Etcheverry</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2019">2019</date>
			<publisher>Iber-LEF@SEPLN</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Deep contextualized word representations</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">E</forename><surname>Peters</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Neumann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Iyyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gardner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Clark</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</author>
		<idno>CoRR abs/1802.05365</idno>
		<ptr target="http://arxiv.org/abs/1802.05365" />
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Overview of the EmoEvalEs task on emotion detection for Spanish at IberLEF 2021</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">M</forename><surname>Plaza-Del-Arco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Jiménez-Zafra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Montejo-Ráez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">D</forename><surname>Molina-González</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">A</forename><surname>Ureña-López</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Martín-Valdivia</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Procesamiento del Lenguaje Natural</title>
		<imprint>
			<biblScope unit="volume">67</biblScope>
			<biblScope unit="issue">0</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">EmoEvent: A Multilingual Emotion Corpus based on different Events</title>
		<author>
			<persName><forename type="first">F</forename><surname>Plaza-Del-Arco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Strapparava</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">A</forename><surname>Ureña-Lopez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Martin-Valdivia</surname></persName>
		</author>
		<ptr target="https://www.aclweb.org/anthology/2020.lrec-1.186" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 12th Language Resources and Evaluation Conference</title>
				<meeting>the 12th Language Resources and Evaluation Conference<address><addrLine>Marseille, France</addrLine></address></meeting>
		<imprint>
			<publisher>European Language Resources Association</publisher>
			<date type="published" when="2020-05">May 2020</date>
			<biblScope unit="page" from="1492" to="1498" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">Improving language understanding by generative pre-training</title>
		<author>
			<persName><forename type="first">A</forename><surname>Radford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Narasimhan</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Effective LSTMs for target-dependent sentiment classification</title>
		<author>
			<persName><forename type="first">D</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Qin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Feng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Liu</surname></persName>
		</author>
		<ptr target="https://www.aclweb.org/anthology/C16-1311" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers</title>
				<meeting>COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers<address><addrLine>Osaka, Japan</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016-12">Dec 2016</date>
			<biblScope unit="page" from="3298" to="3307" />
		</imprint>
	</monogr>
	<note>The COL-ING 2016 Organizing Committee</note>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">DCU: Aspect-based polarity classification for SemEval task 4</title>
		<author>
			<persName><forename type="first">J</forename><surname>Wagner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Arora</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Cortes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Barman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Bogdanova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Foster</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Tounsi</surname></persName>
		</author>
		<idno type="DOI">10.3115/v1/S14-2036</idno>
		<ptr target="https://www.aclweb.org/anthology/S14-2036" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 8th International Workshop on Semantic Evaluation (Se-mEval</title>
				<meeting>the 8th International Workshop on Semantic Evaluation (Se-mEval<address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2014-08">2014. Aug 2014</date>
			<biblScope unit="page" from="223" to="229" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">TDParse: Multi-target-specific sentiment recognition on Twitter</title>
		<author>
			<persName><forename type="first">B</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Liakata</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zubiaga</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Procter</surname></persName>
		</author>
		<ptr target="https://www.aclweb.org/anthology/E17-1046" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics</title>
		<title level="s">Long Papers</title>
		<meeting>the 15th Conference of the European Chapter of the Association for Computational Linguistics<address><addrLine>Valencia, Spain</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2017-04">Apr 2017</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="483" to="493" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Attention-based LSTM for aspect-level sentiment classification</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zhao</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/D16-1058</idno>
		<ptr target="https://www.aclweb.org/anthology/D16-1058" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing</title>
				<meeting>the 2016 Conference on Empirical Methods in Natural Language Processing<address><addrLine>Austin, Texas</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2016-11">Nov 2016</date>
			<biblScope unit="page" from="606" to="615" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
