<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>V. G. Morales-Murillo);</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>LKE-IIMAS team at Rest-Mex 2023: Sentiment Analysis on Mexican Tourism Reviews using Transformer-Based Domain Adaptation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Victor Giovanni Morales-Murillo</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Helena Gómez-Adorno</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>David Pinto</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ilyan Alexey Cortés-Miranda</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pierre Delice</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Benemérita Universidad Autónoma de Puebla, Facultad de Ciencias de la Computación, Language &amp; Knowledge Engineering Lab (LKE)</institution>
          ,
          <addr-line>Puebla</addr-line>
          ,
          <country country="MX">México</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Universidad Nacional Autónoma de México, Facultad de Ciencias</institution>
          ,
          <addr-line>Ciudad de México</addr-line>
          ,
          <country country="MX">México</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Universidad Nacional Autónoma de México, Instituto de Investigaciones en Matemáticas Aplicadas y en Sistemas (IIMAS)</institution>
          ,
          <addr-line>Ciudad de México</addr-line>
          ,
          <country country="MX">México</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2023</year>
      </pub-date>
      <volume>000</volume>
      <fpage>0</fpage>
      <lpage>0002</lpage>
      <abstract>
        <p>Sentiment analysis in tourist texts has gained relevance in the last decade because tourism is a social, cultural, and economic phenomenon that contributes to the economic development of countries. Thus, there is a need to create new natural language processing (NLP) mechanisms that improve tourism services and meet tourist needs. For this reason, LKE-IIMAS team participated in the sentiment analysis task of Rest-Mex 2023, which has three sub-tasks: identifying polarity, type, and country on a given opinion. For solving both sub-tasks, we first generated several training sets using stratified sampling on the training set of this shared task. Then, we applied a back-translation method as a data augmentation technique to tackle the data imbalance in the polarity sub-task. Furthermore, We used a masked language model based on RoBERTa that had been trained with a large Spanish dataset to generate three training strategies. Our ifrst strategy was to fine-tune the base model for text classification. The second strategy was to adapt the base model to the tourism domain and to fine-tune the model adapted to text classification. The third strategy was to fine-tune the model adapted with the data augmented. Finally, our team LKE-IIMAS achieved first in the ranking on the sentiment analysis task at Rest-Mex 2023.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Sentiment analysis</kwd>
        <kwd>Mexican tourism</kwd>
        <kwd>Transformers</kwd>
        <kwd>Domain adaptation</kwd>
        <kwd>NLP</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        This paper describes the participation of the LKE-IIMAS team in the sentiment analysis task
at Rest-Mex 2023 [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. The sentiment analysis in tourist texts has gained relevance in the last
decade because tourism is a social, cultural, and economic phenomenon that contributes to
the economic development of countries [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. For example, this activity in Mexico represents
8.7% of the national gross domestic product (GDP), generating around 4.5 million direct jobs [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ].
Thus, new mechanisms of natural language processing (NLP) are required to improve tourism
services and meet tourist needs. This task is focused on the Spanish language because the most
significant scientific works have focused on the English language [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], and some studies have
focused only on peninsular Spanish. This task provides a text collection directly collected from
tourists’ opinions [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>
        The sentiment analysis task of Rest-Mex 2023 consists in a classification task where the
participating system is asked to predict the polarity of an opinion issued by a tourist who traveled
to the most representative places, restaurants, and hotels in Mexico, Cuba, and Colombia. This
collection was obtained from the tourists who shared their opinion on TripAdvisor between
2002 and 2022. Each opinion’s class is an integer between [
        <xref ref-type="bibr" rid="ref1 ref5">1, 5</xref>
        ], where 1 represents the most
negative polarity, and 5 is the most positive. Also, each opinion has a type label. The problem
is defined as: "Given an opinion about a Mexican tourist place, the goal is to determine the
polarity, between 1 and 5, of the text, the type of opinion (hotel, restaurant or attraction), and
the country of the place of which the opinion is being given (Mexico, Cuba, Colombia)" [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
      </p>
      <p>
        Languages complexity hindered the classification of ironic, sarcastic, and subjective text on
sentiment analysis [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. Additionally, the models based on Long-short term memory (LSTM),
approach fastText, convolutional neural networks (CNN), and transformers have been used as
machine learning (ML) techniques in the literature [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] where the approach fastText has been
faster than other ML models. The transformers model has achieved higher accuracy than other
ML models. Besides, the best results in the previous editions of Rest-Mex were achieved by
transformers approaches [
        <xref ref-type="bibr" rid="ref3 ref7 ref8">3, 7, 8</xref>
        ].
      </p>
      <p>For this reason, we propose a transformer-based domain adaptation to improve sentiment
classification accuracy in tourism reviews. This document is organized as follows. In section 2,
we introduce the related works. In section 3, we describe a background on transformers. In
section 4, we analyze the data. In section 5, we explain the proposed methodology in detail. In
section 6, we show the results, and in section 7, we present our conclusions.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>Rest-Mex’s previous editions encompassed a diverse range of methods for sentiment analysis
classification. These methods included transformer-based models such as fine-tuned versions
of Bert-like models, logistic regression classifiers, the Naive Bayes Multinomial algorithm, a
cascade of binary classifiers, Bayesian techniques, KNN with Jaccard Distance, and an
LDAbased approach for topic extraction. Additionally, a simple Deep Learning architecture was
employed for classification.</p>
      <p>
        In Rest-Mex 2021 [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], the team that achieved the best-performing approach employed two
Bert-based strategies [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. The initial strategy involved fine-tuning BETO, a pre-trained Bert-like
model in Spanish. The second strategy combined Bert embeddings with TF-IDF weighted
feature vectors. The second-best approach in 2021 involved utilizing the BETO model and a
cascade of binary classifiers based on it [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. The third-bet approach proposed an unsupervised
keyword extraction technique to construct a sentiment weight list [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. They employed SVM
for classification, utilizing vector representations of text entities and matching the scores of
prototypical words with the labels of the texts.
      </p>
      <p>For the Rest-Mex 2022 evaluation campaign, the organizers used a new evaluation metric to
assess that particular edition’s sentiment analysis task. This metric is defined in equation 1.
1
measure = 1+MAE +  (1)
2</p>
      <p>Where  is the average among the micro  -measure for each class, and MAE (see equation 2)
evaluates the polarity.</p>
      <p>MAE = 1 ∑︁ | () − ()|
 =1
(2)</p>
      <p>Where  is a participating system ,  () is the result of the instance  according to the
Ground Truth, and () is the output of the participant system , for example, . Finally,  is
the number of instances in the collection.</p>
      <p>
        In Rest-Mex 2022 [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], the team UMU achieved the best results by approaching sentiment
analysis as two distinct challenges: polarity classification as a regression problem and opinion
target resolution as a multi-classification problem. Their pipeline included tasks such as text
cleaning, extracting linguistic features, and utilizing non-contextual and contextual sentence
embeddings using FastText, BERT, and RoBERTa models. They also performed fine-tuning and
hyperparameter optimization, training neural network models with RMSE loss for regression
and binary cross-entropy loss for classification [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. The team that obtained second place,
UC3M, explored two approaches. The first approach involved using SVM, a traditional machine
learning technique, for text classification. The second approach utilized fine-tuned Transformers
with diferent pre-trained models [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. The third-place team, CIMAT, followed a methodology
with two steps. They obtained high-level features from texts using an ensemble of BERT models
trained for both subtasks. In the second step, they created an optimal ensemble of classifiers
trained on the features obtained in the first step to make final predictions for each subtask.
Their preprocessing involved minimal steps, including concatenating the title and opinion and
converting the text to lowercase [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ].
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Background</title>
      <p>
        Transformers became very popular in the Natural Language Processing (NLP) community,
proposed by Vaswani et al. in 2017 [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], this neural architecture achieved and exceeded art
state results for many NLP tasks. Its core mechanism is a self-attention process emphasizing
the contextual relationships of words or tokens.
      </p>
      <p>Transformers addressed the issue of long-term dependencies by utilizing a self-attention
mechanism, which allows the model to weigh the importance of diferent words or tokens in
the sequence while processing the entire sequence simultaneously. By self-attention
mechanism, also known as scaled dot-product attention, the authors allow the model to compute
the importance or attention weights for each word/token in the input sequence based on its
relationships with other words/tokens. The attention allows the model to focus on relevant
input parts during the encoding and decoding stages.</p>
      <p>In this perspective, the use of self-attention mechanisms in Transformers demonstrates
having the capability to perform well in situations of dependencies in the input space, both
locally and on long-range dependencies. This capability has made Transformers highly efective
in various NLP tasks, including sentiment analysis, where capturing contextual information
and dependencies is crucial for accurate sentiment classification. However, some variants of
Transformers are worth emphasizing in the context of our task: Bidirectional Transformers
and Autoregressive Transformers, which are two diferent approaches within the Transformer
architecture that serve diferent purposes.</p>
      <p>
        The Bidirectional Transformers, such as BERT (Bidirectional Encoder Representations from
Transformers), are designed to capture contextual information by considering both the left and
right context of each word in the input sequence. They achieve this through masked language
modeling, where some words in the input are randomly masked, and the model is trained to
predict those masked words based on the surrounding context. BERT-based models have been
widely used for various NLP tasks, including sentiment analysis, as they can efectively leverage
the bidirectional context to understand the relationships between words [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
      </p>
      <p>
        On the other hand, autoregressive Transformers, such as GPT (Generative Pre-trained
Transformer), focus on generating coherent and contextually relevant output by sequentially
processing the input sequence, typically from left to right. The model predicts the next word in the
sequence during training based on the preceding context. This sequential generation approach
makes autoregressive models suitable for text generation and completion tasks. In sentiment
analysis, autoregressive models can generate sentiment-related text, summarize sentiments, or
generate responses based on given sentiments [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
      </p>
      <p>
        Contextual facts are an important piece of information for the sentiment analysis task.
According to the literature [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ], the traditional approach based on bag-of-words or n-gram
ignores the correlation structure between words by treating each word in an isolated manner.
In other words, they ignore the dependency. However, the self-attention mechanism of
transformer architecture considers the importance of context around each word. The self-attention
mechanism allows this architecture to capture long-range dependencies and understand the
sentiment expressed more nuancedly.
      </p>
      <p>There is a wide variety of situations where transformers can improve sentiment analysis. They
learn embeddings for each word in an unsupervised manner. These word embeddings encode
semantic and syntactic information, enabling the model to capture more nuanced sentiment
patterns and generalize better to unseen data.</p>
      <p>For transfer learning, transformers can be pre-trained on large text corpora using tasks like
masked language modeling or next-sentence prediction. This pre-training allows transformers
to learn a general understanding of language, which can then be fine-tuned on specific sentiment
analysis tasks with smaller labeled datasets. Transfer learning helps models leverage knowledge
from diverse text sources, resulting in improved performance on sentiment analysis.</p>
      <p>
        For this reason, the RoBERTa (Robustly Optimized BERT Pretraining Approach) [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] is used by
our team due to this model is a variant of BERT. Furthermore, RoBERTa was trained on a much
larger dataset with 160GB of text, which is more than 10 times larger than the dataset used to
train BERT. On the other hand, RoBERTa has used a more efective training procedure because
it utilizes a dynamic masking technique during training that helps the model learn more robust
and generalizable representations of words.
      </p>
    </sec>
    <sec id="sec-4">
      <title>4. Data Analysis</title>
      <p>The Rest-Mex 2023 dataset contains 359,565 instances from tourists who shared their opinion
on TripAdvisor between 2002 and 2022. This shared task dives its dataset into training and
testing sets. The training set contains 251,702 instances which represent 70% of the original
dataset, while the testing set contains 107,863 instances representing the remaining 30%. Each
instance of the training set has a title, a review which is the opinion issued by the tourist, the
polarity of the opinion (1, 2, 3, 4, 5), the country of the visited place (Mexico, Cuba, Colombia)
and the type of place of which the opinion is being issued (Hotel, Restaurant, Attractive). At
the same time, each instance of the testing set has an identifier, a title, and a review.</p>
      <p>We analyze and process the training set of Rest-Mex 2023, which we divide into training
and testing sets. The training dataset contains 226,531 instances that represent 90% of the
training set of Rest-Mex 2023, while the test dataset contains 25,171 instances representing the
remaining 10%. Figures 1, 2 and 3 present the three kinds of distribution of the training dataset:
polarity, type, and country. It can be observed that this dataset is unbalanced, mainly observed
in the polarity distribution. The majority class represented by label 5 of the polarity class has
the 62.41% of the instances, whereas the minority class represented by label 1 only has the
2.29%. Further, the negative polarity values (1, 2) and the neutral polarity value (3) have a low
number of instances in contrast to positive polarities (values 4 and 5). In the type distribution,
the majority class is the Attractive label with the 44.17% of the instances, whereas the minority
class is the Restaurant label with the 25.61%. In the country distribution, the majority class is
the Mexico label with the 47.18% of the instances, whereas the minority class is the Cuba label
with the 26.31%.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Methodology</title>
      <p>This section describes our methodology to tackle the sentiment analysis task at Rest-Mex 2023.
Figure 4 presents the methodological design composed of 3 main steps: data augmentation,
domain adaptation, and training strategies. In the first step, we use the back-translation
technique to tackle the data imbalance problem. In the second step, we perform a fine-tuning
on a pre-trained language model of RoBERTa to adapt its data in a tourism domain. In the third
step, we design three training strategies for a RoBERTa model based on the Spanish language.
These steps are explained in more detail below.</p>
      <sec id="sec-5-1">
        <title>5.1. Data Augmentation</title>
        <p>
          Data imbalance in Machine Learning refers to an unequal distribution of dataset classes, and
this problem is a challenge that needs more attention to resolve [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ]. There are two main ways
to tackle this problem: the undersampling and the oversampling methods. Undersampling
techniques decrease the number of majority class instances, whereas oversampling methods
increase the number of minority class instances creating new examples or repeating some.
        </p>
        <p>Back-translation is an oversampling technique that translates a text from the target language to
a source language [19]. There are several transformers open source models to perform automatic
text translations.</p>
        <p>We used an oversampling technique based on back-translation with transformers to generate
new reviews by each instance with negative or neutral polarity values. We translate the text
review of each instance of the training dataset to a target language, and the result is translated
again to its source language, Spanish. So, several languages are employed as target languages.
Three new instances are generated by each instance with label 1 using the English, French,
and Deutsch languages as target languages. We generate two new instances by each instance
with label 2 using the English and French languages as target languages. We produce one new
instance by each instance with label 3 using as the target language the English language. Table 1
shows the transformers translation models used to perform back-translation of minority classes
on the training dataset, the source language and the target language used by these models, and
the labels of the instances translated by these models.</p>
        <p>Figure 5 shows the new distribution of the polarity class in the training dataset, where 48,118
new instances are included and assigned to labels 1, 2, and 3. We generate a dataset with this
new instances called analisis-sentimientos-textos-turisitcos-mx-polaridadV3-DA.</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Domain Adaptation</title>
        <p>
          Transformers models are used in many applications of NLP because they provide good results
with transfer learning when the corpus used for pretraining a model is not too diferent from
the corpus used for fine-tuning the specific task at hand. However, there are some cases when
you want to fine-tune the language models on your data before training a task-specific head.
So, this process of fine-tuning a pre-trained language model on in-domain data is called domain
adaptation [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. Such as, a legal contracts dataset hards the transfer learning with a transformer
model as BERT because this model will process the domain-specific words of this dataset as rare
tokens, and the performance results may not be satisfactory. For this reason, domain adaptation
is an alternative for this problem, and it can increase the performance of many post tasks. It
usually is achieved only one time [20].
        </p>
        <p>
          We fine-tune a pre-trained model called RoBERTa-base-bne [21] that has been trained with a
large corpus of Spanish with 570GB of clean and deduplicated text processed, which was gotten
from a web crawling performed by the National Library of Spain from 2009 to 2019. We use an
unsupervised dataset with 63,740 instances [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. We join this unsupervised dataset with 251,702
instances of the training set of Rest-Mex 2023 to produce 315,442 instances of a new dataset that
contains only text reviews without polarity, type, and country labels. In the preprocessing step,
all instances’ texts of the dataset are tokenized, concatenated, and divided into chunks of equal
size to avoid that individual text might get truncated if their text is too long. Therefore, this step
can reduce the losing information. The preprocessing result is a dataset with 188,768 tokenized
instances with input_ids, attention_mask, word_ids, and labels as features. In fine-tuning, we
randomly mask the 15% of the tokens in each batch of text as in the following example.
• &lt;s&gt;Me ha parecido un museo excepcional,&lt;mask&gt; en una zona&lt;mask&gt; accesible de
la Ciudad de México. Las salas son impresionantes carpintería, sobre todo las &lt;mask&gt;
Teotihu&lt;mask&gt;án y la de la cultura Maya&lt;mask&gt; Existe la posibilidad de que te acompañe
un especialista&lt;mask&gt; la visita y&lt;mask&gt;&lt;mask&gt; pena. &lt;mask&gt;. . . &lt;/s&gt;;
        </p>
        <p>As you can see, the &lt;mask&gt; token has been randomly inserted at various locations of text
and our model should predict these tokens during training process. Furthermore, we divide the
dataset preprocessed into training and testing sets, where the training datset contents 169,892
instances and the test dataset contens 18,876 instances. The main arguments for the training
process are showed at Table 2.</p>
        <p>
          Perplexity [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] is a standard metric to evaluate the performance of language models due to that
masked language models don’t use datasets labeled, so this metric calculates the probabilities it
assigns to the next word in all the sentences of the test dataset where high probabilities mean
that model is not perplexed by the testing instances. We use the exponential of the cross-entropy
loss as a mathematical definition of perplexity. We calculate the perplexity by evaluating the
Transformers function, where a lower perplexity score means a better language model. The
perplexity of our model without running the training is 28.03. In contrast, the perplexity of
our model trained is 6.05, which is lower than the first result, representing good results for
the domain adaptation on RoBERTa-base-bne. The masked language model result is called
roberta-base-bne-finetuned-TripAdvisorDomainAdaptation.
        </p>
      </sec>
      <sec id="sec-5-3">
        <title>5.3. Training strategies</title>
        <p>Three main strategies for training are designed using a masked language model called
RoBERTabase-bne [21] that has been pre-trained using a large Spanish corpus with 570GB of clean and
deduplicated text. Furthermore, the texts were obtained with a web crawling performed by the
National Library of Spain from 2009 to 2019. We explain the strategies below.</p>
        <p>
          In the first strategy, we fine-tune RoBERTa-base-bne for classifying polarity, type, and country
on TripAdvisor review texts. We use of training set of Rest-Mex 2023 and divide it in a manner
stratified by each class into training and test sets. The training dataset contains 70% of the
training set instances of Rest-Mex 2023, whereas the test dataset contains the remaining 30%.
Besides, we join the title and review of the instances as the only feature, rename the target class
as labels, and remove the rest of the features. The result of this process is three datasets called
analisis-sentimeinto-textos-turisitcos-mx-polaridad,
analisis-sentimeinto-textos-turisitcos-mxtipo and analisis-sentimeinto-textos-turisitcos-mx-pais (see https://huggingface.co/alexcom/).
Then for the classification with Transformers, the polarity labels are modified by our team
from [
          <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4 ref5">1, 2, 3, 4, 5</xref>
          ] to [
          <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4">0, 1, 2, 3, 4</xref>
          ], so the type labels are switched from [Attractive, Hotel,
Restaurant] to [
          <xref ref-type="bibr" rid="ref1 ref2">0, 1, 2</xref>
          ] and the country labels are changed from [Mexico, Colombia, Cuba] to [
          <xref ref-type="bibr" rid="ref1 ref2">0,
1, 2</xref>
          ]. We tokenize each dataset with the RoBERTa-base-bne AutoTokenizer of Transformers to
generate the features of labels, input_ids, and attention_mask. Moreover, we select the number
of labels by each class. For example, the number of the labels to polarity classification is 5
and the number of the labels to type and country classification is 3. Besides, we use F1 metric
for the model evaluation and show some training parameters for text classification in Table 3.
Also, we use only 20,000 instances of the training dataset. We produce with this process three
models called roberta-base-bne-finetuned-analisis-sentimiento-textos-turisticos-mx-polaridad
to polarity classification,
roberta-base-bne-finetuned-analisis-sentimiento-textos-turisticos-mxtipo to type classification and
roberta-base-bne-finetuned-analisis-sentimiento-textos-turisticosmx-pais (see https://huggingface.co/vg055/) to country classification.
        </p>
        <p>In the second strategy, we ifne-tune
roberta-base-bne-finetunedTripAdvisorDomainAdaptation to classify polarity, type, and country on TripAdvisor
review texts. We use of training set of Rest-Mex 2023 and divide it in a manner stratified
by each class into training and test sets. The training dataset contains 90% of training
set instances of Rest-Mex 2023, whereas the test dataset contains the remaining 10%. We
process the dataset of same way that the first strategy and we generate three datasets called
analisis-sentimientos-textos-turisitcos-mx-polaridadV2,
analisis-sentimientos-textos-turisitcosmx-tipoV2 and analisis-sentimientos-textos-turisitcos-mx-paisV2. Then, we perform the
same steps from first strategy on the tokenizing, preprocessing and training but now with
the model adapted to tourism domain and we now use all instances of training dataset for
the training. We produce with this process three models called
roberta-base-bne-finetunedTripAdvisorDomainAdaptation-finetuned-e2-RestMex2023-polaridad to polarity classification,
roberta-base-bne-finetuned-TripAdvisorDomainAdaptation-finetuned-e2-RestMex2023-tipo
to type classification and
roberta-base-bne-finetuned-TripAdvisorDomainAdaptation-finetunede2-RestMex2023-pais to county classification.</p>
        <p>In the third strategy, we fine-tune roberta-base-bne-finetuned-TripAdvisorDomainAdaptation
to classify polarity on TripAdvisor review texts. Unlike the other strategies, we use the dataset
analisis-sentimientos-textos-turisitcos-mx-polaridadV3-DA, which is based on the training
set of Rest-Mex 2023. Still, the training instances of negative and neutral classes were
increased by a data augmentation technique applied previously. We repeat the process of the
second strategy with this dataset and generate one model called
roberta-base-bne-finetunedTripAdvisorDomainAdaptation-finetuned-e2-RestMex2023-polaridadDA-V3 to polarity
classification.</p>
        <p>We performed in total 23 experiments, 19 experiments to polarity identification, 2 experiments
to type classification and 2 experiments to country identification. We focused on polarity
identification due to the problem is more complex for several reasons as the data imbalance
and the classification of 5 classes. We included only the most representative results of these
experiments by each training strategy on the next section.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Results</title>
      <p>The results obtained during the development period are presented in Figure 6 where the second
strategy presents the best result, followed by the third and first strategies. Polarity, type,
and country classification results related to the first strategy are improved with the domain
adaptation applied to the second strategy. However, the data augmentation on the third strategy
does not represent an improvement to the second strategy, whereby it is essential to review the
data augmentation techniques in more detail for future work.</p>
      <p>We generate three outputs based on these strategies and the test set of Rest-Mex 2023. All
outputs use the second strategy for the classification of type and country. Output 1 utilizes
the first strategy for polarity classification, output 2 employs the second strategy for polarity
classification, and output 3 applies the third strategy to classify the polarity. We send these
outputs to Rest-Mex 2023, and the results gotten on the test process of Rest-Mex 2023 are
presented in Figure 7, where the best result is achieved by output 2, followed by output 1 and
output 3. However, the diference between the three outputs is very low due to only the polarity
classification diferentiates these outputs. On the other hand, we observe that domain adaptation
is an excellent alternative to improve our results. However, looking for other methods to tackle
the unbalance problem of the polarity classification is essential.</p>
      <p>We achieved first place in the ranking with output 2 on the sentiment analysis task of
RestMex 2023, where 18 teams from around the world participated. Besides, our three outputs were
ranked top three in this shared task.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusions</title>
      <p>Our team, called LKE-IIMAS, achieve the highest results on the sentiment analysis task at
Rest-Mex 2023 with a domain adaptation to a pre-trained model based on RoBERTa trained with
a large dataset in the Spanish language. We developed three main steps in our methodology:
data augmentation, domain adaptation, and training strategies. Besides, we generated several
datasets with diferent types of stratification by classifying polarity, type, and country. We
finetuned a masked language model to apply the domain adaptation and evaluated their perplexity.
We fine-tuned a language model for polarity, type, and country text classification. We evaluated
the result of these models with the F1 metric to select the three best strategies to generate three
output submissions to Rest-Mex 2023. Domain adaptation improved the results. However, it
is important to analyze other data augmentation methods that produce new instances with
quality.</p>
    </sec>
    <sec id="sec-8">
      <title>Acknowledgments</title>
      <p>This work has been carried out with the support of DGAPA-UNAM PAPIIT project number
TA101722. The authors also thank the Language &amp; Knowledge Engineering Lab (LKE) of
Benemérita Universidad Autónoma de Puebla, Instituto de Investigaciones en Matemáticas
Aplicadas y en Sistemas (IIMAS) of Universidad Nacional Autónoma de México and CONACYT
for the computing resources provided through the Deep Learning Platform for Language
Technologies of the INAOE Supercomputing Laboratory. We also want to thank Eng. Roman
Osorio for supporting the student administration of the project.
243–248. doi:10.1109/ICICS49469.2020.239556.
[19] N. L. Pham, V. Vinh Nguyen, T. V. Pham, A data augmentation method for
englishvietnamese neural machine translation, IEEE Access 11 (2023) 28034–28044. doi:10.1109/
ACCESS.2023.3252898.
[20] Y. Zhu, Y. Qiu, Q. Wu, F. L. Wang, Y. Rao, Topic driven adaptive network for cross-domain
sentiment classification, INFORMATION PROCESSING &amp; MANAGEMENT 60 (2023).
doi:10.1016/j.ipm.2022.103230.
[21] A. Gutiérrez-Fandiño, J. Armengol-Estapé, M. Pàmies, J. Llop-Palao, J. Silveira-Ocampo,
C. P. Carrino, A. Gonzalez-Agirre, C. Armentano-Oller, C. Rodriguez-Penagos, M. Villegas,
Spanish language models, 2021. arXiv:2107.07253.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <article-title>Álvarez-Carmona, Á</article-title>
          .
          <string-name>
            <surname>Díaz-Pacheco</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Ramón</surname>
            ,
            <given-names>A. Y.</given-names>
          </string-name>
          <string-name>
            <surname>Rodríguez-González</surname>
          </string-name>
          , L. BustioMartínez, V.
          <string-name>
            <surname>Muñis-Sánchez</surname>
            ,
            <given-names>A. P.</given-names>
          </string-name>
          <string-name>
            <surname>Pastor-López</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Sánchez-Vega</surname>
          </string-name>
          ,
          <source>Overview of RestMex at IberLEF 2023: Research on Sentiment Analysis Task for Mexican Tourist Texts, Procesamiento del Lenguaje Natural</source>
          <volume>71</volume>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Diaz-Pacheco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Álvarez</surname>
          </string-name>
          <string-name>
            <surname>Carmona</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. A. C.</given-names>
            <surname>Chávez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Ramírez-Silva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <article-title>Artificial intelligence methods to support the research of destination image in tourism. a systematic review</article-title>
          ,
          <source>Journal of Experimental &amp; Theoretical Artificial Intelligence</source>
          <volume>0</volume>
          (
          <year>2022</year>
          )
          <fpage>1</fpage>
          -
          <lpage>31</lpage>
          . URL: https://doi.org/10.1080/0952813X.
          <year>2022</year>
          .
          <volume>2153276</volume>
          . doi:
          <volume>10</volume>
          .1080/0952813X.
          <year>2022</year>
          .
          <volume>2153276</volume>
          . arXiv:https://doi.org/10.1080/0952813X.
          <year>2022</year>
          .
          <volume>2153276</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Álvarez</surname>
          </string-name>
          <string-name>
            <surname>Carmona</surname>
          </string-name>
          , Ángel Díaz-Pacheco,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Fajardo-Delgado</surname>
          </string-name>
          , R. G.
          <string-name>
            <surname>-R. y Lázaro</surname>
          </string-name>
          Bustio-Martínez,
          <article-title>Overview of rest-mex at iberlef 2022: Recommendation system, sentiment analysis and covid semaphore prediction for mexican tourist texts</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>69</volume>
          (
          <year>2022</year>
          )
          <fpage>289</fpage>
          -
          <lpage>299</lpage>
          . URL: http://journal.sepln.org/sepln/ojs/ojs/index.php/pln/article/view/6449.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Álvarez</surname>
          </string-name>
          <string-name>
            <surname>Carmona</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-Gonzalez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Fajardo-Delgado</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. G.</given-names>
            <surname>Sánchez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Pérez-Espinosa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Martínez-Miranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodríguez</surname>
          </string-name>
          , L. BustioMartínez, Ángel Díaz-Pacheco,
          <article-title>Natural language processing applied to tourism research: A systematic review and future research directions</article-title>
          ,
          <source>Journal of King Saud University - Computer and Information Sciences</source>
          <volume>34</volume>
          (
          <year>2022</year>
          )
          <fpage>10125</fpage>
          -
          <lpage>10144</lpage>
          . URL: https: //www.sciencedirect.com/science/article/pii/S1319157822003615. doi:https://doi.org/ 10.1016/j.jksuci.
          <year>2022</year>
          .
          <volume>10</volume>
          .010.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>CIMAT</surname>
          </string-name>
          , Rest-Mex
          <source>: Research on Sentiment Analysis Task for Mexican Tourist Texts</source>
          ,
          <year>2023</year>
          . URL: https://sites.google.com/cimat.mx/rest-mex2023/.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>S.</given-names>
            <surname>Balcí</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. M.</given-names>
            <surname>Demirci</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Demirhan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sarp</surname>
          </string-name>
          ,
          <article-title>Sentiment Analysis Using State of the Art Machine Learning Techniques</article-title>
          , in: C.
          <string-name>
            <surname>Biele</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Kacprzyk</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          <string-name>
            <surname>Kopeć</surname>
            ,
            <given-names>J. W.</given-names>
          </string-name>
          <string-name>
            <surname>Owsiński</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Romanowski</surname>
          </string-name>
          , M. Sikorski (Eds.),
          <source>Proceedings of MIDI'2021 - 9th Machine Intelligence and Digital Interaction Conference</source>
          , Springer International Publishing, Warsaw,
          <year>2022</year>
          , pp.
          <fpage>34</fpage>
          -
          <lpage>42</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Álvarez-Carmona</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Aranda</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Arce-Cardenas</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Fajardo-Delgado</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>GuerreroRodríguez</surname>
            ,
            <given-names>A. P.</given-names>
          </string-name>
          <string-name>
            <surname>López-Monroy</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Martínez-Miranda</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Pérez-Espinosa</surname>
            ,
            <given-names>A. Y.</given-names>
          </string-name>
          <article-title>RodríguezGonzález, Overview of rest-mex at iberlef 2021: Recommendation system for text mexican tourism</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Álvarez-Carmona</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Aranda</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Guerrero-Rodríguez</surname>
            ,
            <given-names>A. Y.</given-names>
          </string-name>
          <string-name>
            <surname>Rodríguez-González</surname>
            ,
            <given-names>A. P.</given-names>
          </string-name>
          <string-name>
            <surname>López-Monroy</surname>
          </string-name>
          ,
          <article-title>A combination of sentiment analysis systems for the study of online travel reviews: Many heads are better than one</article-title>
          ,
          <source>Computación y Sistemas</source>
          <volume>26</volume>
          (
          <year>2022</year>
          ). doi:https://doi.org/10.13053/CyS-26-2-4055.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J.</given-names>
            <surname>Vásquez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Gómez-Adorno</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Bel-Enguix</surname>
          </string-name>
          ,
          <article-title>Bert-based approach for sentiment analysis of spanish reviews from tripadvisor</article-title>
          ,
          <source>CEUR WS Proceedings</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>163</fpage>
          -
          <lpage>172</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ballester-Espinosa</surname>
          </string-name>
          ,
          <article-title>Cascade of biased two-class classifiers for multi-class sentiment analysis</article-title>
          ., in: IberLEF@SEPLN, CEUR-WS.org,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>Geovanni</given-names>
            <surname>Velazquez</surname>
          </string-name>
          <string-name>
            <surname>Medina</surname>
          </string-name>
          , Delia Irazú Hernández Farías,
          <article-title>Dci-ug participation at rest-mex 2021: A transfer learning approach for sentiment analysis in spanish</article-title>
          ., in: IberLEF@SEPLN, CEUR-WS.org,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>J. A.</given-names>
            <surname>García-Díaz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Rodríguez-García</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>García-Sánchez</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Valencia-García</surname>
          </string-name>
          , Umuteam at REST-MEX
          <year>2022</year>
          :
          <article-title>Polarity prediction using knowledge integration of linguistic features and sentence embeddings based on transformers</article-title>
          , in: M.
          <article-title>Montes-y-</article-title>
          <string-name>
            <surname>Gómez</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Gonzalo</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Rangel</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Casavantes</surname>
            ,
            <given-names>M. Á.</given-names>
          </string-name>
          <string-name>
            <surname>Á. Carmona</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Bel-Enguix</surname>
            ,
            <given-names>H. J.</given-names>
          </string-name>
          <string-name>
            <surname>Escalante</surname>
          </string-name>
          , L. A.
          <string-name>
            <surname>de Freitas</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Miranda-Escalada</surname>
            ,
            <given-names>F. J.</given-names>
          </string-name>
          <string-name>
            <surname>Rodríguez-Sanchez</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Rosá</surname>
            ,
            <given-names>M. A. S.</given-names>
          </string-name>
          <string-name>
            <surname>Cabezudo</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Taulé</surname>
          </string-name>
          , R. Valencia-García (Eds.),
          <source>Proceedings of the Iberian Languages Evaluation Forum (IberLEF</source>
          <year>2022</year>
          )
          <article-title>co-located with the Conference of the Spanish Society for Natural Language Processing (SEPLN</article-title>
          <year>2022</year>
          ),
          <string-name>
            <given-names>A</given-names>
            <surname>Coruña</surname>
          </string-name>
          , Spain,
          <year>September 20</year>
          ,
          <year>2022</year>
          , volume
          <volume>3202</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2022</year>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3202</volume>
          /restmex-paper13.
          <fpage>pdf</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaswani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Parmar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uszkoreit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Gomez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Kaiser</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Polosukhin</surname>
          </string-name>
          ,
          <article-title>Attention is all you need</article-title>
          ,
          <source>in: Proceedings of the 31st International Conference on Neural Information Processing Systems</source>
          , NIPS'17, Curran Associates Inc.,
          <string-name>
            <surname>Red</surname>
            <given-names>Hook</given-names>
          </string-name>
          ,
          <string-name>
            <surname>NY</surname>
          </string-name>
          , USA,
          <year>2017</year>
          , p.
          <fpage>6000</fpage>
          -
          <lpage>6010</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , Bert:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          , ArXiv abs/
          <year>1810</year>
          .04805 (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>A.</given-names>
            <surname>Katharopoulos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Vyas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Pappas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Fleuret</surname>
          </string-name>
          ,
          <article-title>Transformers are rnns: Fast autoregressive transformers with linear attention</article-title>
          ,
          <source>in: International Conference on Machine Learning</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>L.</given-names>
            <surname>Tunstall</surname>
          </string-name>
          , L. von
          <string-name>
            <surname>Werra</surname>
          </string-name>
          ,
          <source>T. Wolf, Natural Language Processing with Transformers</source>
          ,
          <string-name>
            <surname>Revised Edition</surname>
            ,
            <given-names>O</given-names>
          </string-name>
          <string-name>
            <surname>'Reilly Media</surname>
          </string-name>
          ,
          <year>2022</year>
          . URL: https://books.google.com.mx/books?id= 8plxEAAAQBAJ.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Joshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          ,
          <article-title>Roberta: A robustly optimized BERT pretraining approach</article-title>
          , CoRR abs/
          <year>1907</year>
          .11692 (
          <year>2019</year>
          ). URL: http://arxiv.org/abs/
          <year>1907</year>
          .11692. arXiv:
          <year>1907</year>
          .11692.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>R.</given-names>
            <surname>Mohammed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Rawashdeh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Abdullah</surname>
          </string-name>
          ,
          <article-title>Machine learning with oversampling and undersampling techniques: Overview study and experimental results</article-title>
          ,
          <source>in: 2020 11th International Conference on Information and Communication Systems (ICICS)</source>
          ,
          <year>2020</year>
          , pp.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>