<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>UMSNH at RestMex 2023: An XGBoost Stacking with Pre-Trained Word-Embeddings over Data Batches</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jaime Cerda-Flores</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Rodrigo Hernández-Mazariegos</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jesús Ortiz-Bejar</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Félix Calderón-Solorio</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>José Ortiz-Bejar</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>División de Estudios de Posgrado, Faultad de Ingeniería Eléctrica, Universidad Michoana de San Nicolás de Hidalgo</institution>
          ,
          <addr-line>Morelia, Michoacán</addr-line>
          ,
          <country country="MX">México</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Facultad de Ciencias Físico Matemáticas , Universidad Michoana de San Nicolás de Hidalgo</institution>
          ,
          <addr-line>Morelia, Michoacán</addr-line>
          ,
          <country country="MX">México</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The present work resumes the participation of the UMSNH Team at RestMex 2023. Our approach aims to categorize tourist destination reviews in Cuba, Colombia, and Mexico in three diferent aspects: the country to which the attraction belongs, the type of attraction, and the level of tourist satisfaction. For this purpose, a variety of diferent Word-Embeddings (FastText,  -TC, BERT) will be used to generate a baseline, after which combinations of these will be assembled using XGBoost to try to improve on their individual results.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Word-emeddings</kwd>
        <kwd>BERT</kwd>
        <kwd>FastText</kwd>
        <kwd>stacking</kwd>
        <kwd>XGBoost</kwd>
        <kwd>-TC</kwd>
        <kwd>sentiment analysis</kwd>
        <kwd>text categorization</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        which was previously labeled according to a predefined set of categories (training set). Then a
prepossessing step is performed; the latter implies normalizing (case normalization, removing
stop works, steaming, strategies to deal with links, usernames, hashtags, etc). From normalized
documents, a vocabulary is produced through tokenization. A numerical representation for
each model can be built and used from the vocabulary to fit a machine-learning algorithm [
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ].
Transforming documents into a set of numerical vectors (often called Vector Space Model or
VSM) is a challenging task. Fortunately, there is a vast collection of methodologies and research
that deals exclusively with the topic.
      </p>
      <p>Undoubtedly, the representation used directly impacts the performance of the classification
algorithm; in other words, the quality of the model will depend on the VSM. For our approach,
an ensemble of multiple of light and fast pre-trained models is used to produce VSMs, which
are used to train a Multi-Layer Perceptron (MLP) to build and stacked VSM, which is finally
ensembled through XGBoost classifier.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Task Description</title>
      <p>The challenge we took for the RestMex 2023 was the Sentiment Analysis Task. For this
assignment, the objective was to develop a system capable of classifying the polarity of an opinion
given by tourists on diferent types of establishments in Mexico, Cuba, and Colombia. The
system also had to determine the type of establishment the opinion was given on between three
classes (Attractive, Hotel, and Restaurant) and the country in which it is located. Detailed task
and dataset descriptions may be found in the overview document [5].</p>
      <sec id="sec-2-1">
        <title>2.1. Data set</title>
        <p>The training data set and the test data set were given by the organizers of RestMex 2023. The
training data set consisted of an xlsx file with 251,702 sample reviews. Each sample has its
corresponding Title column, Review column, Polarity label column, Country label column, and
Type label column. The Polarity column represents the sentiment expressed in the review on a
numeric scale ranging from 1 to 5, where 1 indicates a highly negative sentiment and a value of
5 represents a highly positive sentiment. There were no missing values for any columns in the
training data set.</p>
        <p>The test data set was also an xlsx file that contained 107,863 samples. Unlike the training
data set, each sample only had an ID column (from 0 to 107862), a Title column, and a Review
column.</p>
        <p>The reviews were collected from TripAdvisor submissions by tourists between 2002 and 2022.</p>
        <p>Table 1 shows the distribution of classes for each feature in the training data set. The
distribution of classes for the test data set is unknown.</p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Evaluation</title>
        <p>The organizers proposed the following equations to evaluate the sentiment score of a system .
.</p>
        <sec id="sec-2-2-1">
          <title>Polarity</title>
        </sec>
        <sec id="sec-2-2-2">
          <title>Country</title>
        </sec>
        <sec id="sec-2-2-3">
          <title>Class</title>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Vector Space Models</title>
      <p>The first step to transforming text into vectors involves building a vocabulary. The goal is
to convert text sequences into minimal meaningful writing units (tokens); for example, split
a sentence into words. As we deal with written language, obtaining the tokens requires the
manipulation of character strings. It is required to identify punctuation, diacritics (dieresis,
accents). In the case of English, you may want to divide the contractions (You’re →− you are).
From the vocabulary, it is possible to build a vector representation.</p>
      <p>Note that tokenization could be at the sufix/prefix level, syllables, or even letters. It is also
possible to build units made up of 2,3 or  words; these tokens are known as -grams and allow
us to include concepts of more than one unit, for example, in English terms like ice cream or
New York.</p>
      <sec id="sec-3-1">
        <title>3.1. Tokenizing and vectorizing</title>
        <p>The tokenization process is a document segmentation process. The segmentation is dividing
the text (unstructured information) into smaller units that can be accounted for discreetly. The
result of counting the occurrences of each term, known as bag of words, can be used directly as
a vector representation of the document. Information retrieval and search are the most common
application of this type of vector (bag of words) for document retrieval or search.</p>
        <p>As an example, consider the following sentence 1 "Rain may fall, and storms may come"
(the sentence was taken from the lyrics of the song ’Bob Lennon’ in the English version of the
manga ’20th Century Boys’ by Urasawa, N. (2000-2006), published by VIZ Media). By tokenizing
1 by words, lower-casing, and removing punctuation, a simple way of obtaining a numerical
representation of a text sequence is employing a binary representation of each token that
exists in the vocabulary; this representation is known as one-hot vectors. Each word may be
represented as a one-hot vector, where each one-hot binary vector will be the vocabulary size
and will have only a 1 in the position that corresponds to the word it represents. Therefore,
each document vector(matrix) is comprised of the list of the tokens in the document (see Table
2).</p>
        <p>In Table 2, each row is the one-hot vector for one word in 1, for instance the token fall is
represented by the vector {0, 0, 1, 0, 0, 0}.</p>
        <p>For our example, a 7 × 6 matrix since the vocabulary is only made up of a single sentence.
In a matrix document, one indicates that the token is part of the document, and zero indicates
that that term is not. This type of representation is often used in neural networks and language
modeling. The importance of any vector representation is that it allows transforming sentences
written in natural language into a space where it is possible to perform mathematical operations
to apply classification models.</p>
        <p>One of the disadvantages of having a matrix representation of one-hot vectors is that they
are highly dispersed; their storage can be ineficient if done in a matrix way, while if we do
it through lists or some other dispersed structure, there is an increase in the complexity of
the operations. For instance, in the Spanish language, it is considered that there are around
100000 words [6]. As a result, a one-hot vector for a word will be a 100000-dimensional vector
with only a few elements diferent from 0. Even with the previous disadvantage, this vector
would preserve the concepts that appear in the document; it would be similar to an index of
terms in a book. Moreover, a binary vector is enough to determine whether or not the word
appears in the document due to it is frequently used in search and information retrieval. A
simple "improvement" for binary one-hot encoding is counting how many times each token
appears in the sentence. The latter can be seen as assigning a weight based on per document
word frequency.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Frequency Vector Spaces</title>
        <p>Even though identifying and counting terms in the vocabulary is helpful for simple problems,
such as computing usage statistics or performing keyword searches, for more complex cases
we might also want to perform tasks in which the importance within the document or the
collection of the words is relevant. For example, it is not desirable for an email to be classified
as spam if it only contains a common term in that type of message.</p>
        <p>There are plenty of weighted models; they aim to provide a representation that captures the
importance of each token. One of the most popular representations is frequency-based, like
TF and TF-IDF (Term Frequency Inverse Document Frequency). TF refers to the frequency of
a term in a document, while IDF implies that TF will be weighed based on the frequency of
that same term in the entire collection. Schemes like TF-IDF help to relate a token to a specific
document in a corpus, assigning a numerical value to the importance of that word in the given
document depending on its frequency within the entire corpus.
4.  -TC
The  -TC system works with frequency vector spaces.  -TC is a versatile and compact tool
created to handle various tasks related to text classification, regardless of the specific domain
or language involved. This is achieved by treating the creation of eficient text classifiers as
a problem of optimizing combinations of function parameters applied over a corpus of text.
The functions involved encompass text modifications that normalize a given text, tokenization
options for the text, and weighting techniques, followed by vectorizing the resulting vocabulary,
which is used to train a Support Vector Machine (SVM). For the approach evaluated in this work,
the output probabilities of the SVM are stacked as the input for an XGB Classifier to perform
new predictions. Even though  -TC is not a word embedding or deep learning approach, it is a
competitive scheme used successfully in multiple text classification contests (for instance [ 7, 8]).
For a detailed description of how  -TC works, the reader may consult [9].</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>5. Word-Embeddings</title>
      <p>In the same way as the frequency-based techniques, word embeddings are used to transform
text into numeric vectors. The main diference is that word embeddings are dense vectors
usually learned by a deep neural network. Each word is assigned a unique vector so that similar
words must have vectors close to each other. Word embeddings are based on the idea that the
words around them influence a word’s meaning. In the following subsection, a brief review of
some of the embeddings we used is given.</p>
      <sec id="sec-4-1">
        <title>5.1. FastText</title>
        <p>FastText is a library developed by the Facebook AI Research lab for text classification and
representation. FastText is based in skipgram and cbow architectures introduced by Mikolov in
[10]. This tool transforms text into continuous vectors, enabling its application to supervised
and unsupervised training tasks. Like with  -TC, it is also essential to find the best combination
of function parameters available for building eficient models, although we used base parameters
for our work. The classification task is done using multinomial logistic regression, taking the
sentence or document vector obtained during the model-building phase as the features. An
interested reader in FastText will find a full description in [11].
5.2. BERT
Bidirectional Encoder Representations from Transformers (BERT), first introduced in [ 12] is a
deep learning approach tool capable of managing several NLP tasks, including text classification.
A transformer is a type of neural network architecture that employs a self-attention mechanism
that allows the model to weigh the importance of diferent words in a sentence to determine their
relationship. BERT processes text using a bidirectional approach. This means BERT can analyze
a word in the context of both preceding and following words, providing a better understanding
of the word’s meaning. It was pre-trained using text from the English Wikipedia and the Toronto
BookCorpus utilizing Masked Language Modeling (MLM) and Next Sentence Prediction (NSP).
In our case, the resulting model is then fine-tuned to a specific task, which is text classification.
Given that the BERT model used was pre-trained on an English database and the samples from
the RestMex data sets were in Spanish, it was necessary to use a multi-lingual model that
generates aligned vector spaces.[13]. You may be referred to [14] for a comprehensive survey
about BERT.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>6. An ensemble word-embedding batched approach</title>
      <p>For the proposed task consisting of a pretty massive data set, a modest computer (Intel(R) Xeon(R)
W-2125 CPU @ 4.00GH, with 32 GB ram) is not enough to process all data at the same time for
most of the pre-trained word-embeddings; Our team used a batched approach where the 251702
documents where divided into  batches using  −  as training set and the remaining  for
model validation. Figure 1 depicts our approach. First, data is split into  equal-size batches,
one  TC, and a FastText model is fitted for each batch; as BERT is not a complete classification
system (we only use the pre-trained model), documents in batches are transformed in BERT
sentences embeddings by averaging all words in each document. A Multi-Layer Perceptron
(MLP) is fitted. A new VSM is built by concatenating the probabilities classes output for each of
the fitted models; the new VSM is used to fit an XGB classifier. The XGB classifier performs
the final prediction. Please note that the input dimension of the data fed into XGBoost is given
by  ×  × , where  represents the number of classes,  denotes the number of training
batches, and  represents the number of used VSMs. In our experiments,  is limited to the
set {1, 2, 3}.</p>
      <p>BERT 0
BERT 1
BERT n
0
1
n
FastText 0
FastText 1
FastText n</p>
      <p>MLP 1
MLP 2
MLP n</p>
      <p>XGB</p>
      <p>Final
prediction
DB
batch 0
batch 1
batch n</p>
      <p>It is worth mentioning that an instance of the pipeline is run for each of the three sub-tasks
proposed for the RestMEX 2023 (i.e., Attraction, Country, and polarity).</p>
      <sec id="sec-5-1">
        <title>6.1. XGBoost</title>
        <p>XGBoost[15] is an optimized distributed gradient boosting open source library designed to
be highly eficient, flexible, and portable. It works under the gradient boosting framework.
Gradient boosting tries to, given two weak classifiers models 1, 2, build a stronger classifier
3 following a Taylor expansion where:
 2 =  (, 1() +  2())
3 =1 +  22
(5)
(6)</p>
        <p>Where  is the predefined lost function. Once we have 3, we can iterate over all our weak
models to improve gradually; note how this is a Taylor expansion of first order; XGBoost uses a
second-order Taylor expansion introducing the gradient and hessian of 1.</p>
      </sec>
      <sec id="sec-5-2">
        <title>6.2. Experimental setup</title>
        <p>For our experiments, the number of batches  was fixed to 20, where four (  = 4) are separated
as a validation set and 16 for training the model; a cross-validation with five folds is performed
to measure the generalization capacity for our model. For the FastText and  TC classifier,
the default system parameters are kept and applied to each of the 20 batches. For  TC, no
prepossessing is used over the text because it is one of the parameters optimized by  TC. On
the other hand, for BERT and FastText, the only preprocessing applied is normalizing the text
to lowercase. Please note that while FastText and  TC predictions are fed forwardly to XGB,
BERT is used to produce document embeddings by averaging the vector representation for the
words in each text; the embeddings are used to train an MLP, which outputs are the inputs used
to the staking process.</p>
      </sec>
      <sec id="sec-5-3">
        <title>6.3. Results</title>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>7. Conclusions and future work</title>
      <p>From the results, it is clear that assembling improves on the results given by the individual
embeddings. However, it is noticeable that when using a combination of the three models,
there is a slight decrease in the score for Sentiment and Polarity, which indicates that doing so
generates a little loss of information. Given the results from the RestMex competition, where
our work placed us in the middle of the other teams, the presented methodology could be
considered as a baseline for diferent events as it is relatively lightweight and computationally
undemanding.</p>
      <p>As for future work, we would like to experiment with diferent sizes of batches to see if this
afects the results somehow, as well as improving the models generated by FastText thorugh
parameter optimization. We can now use heavier embedding methods (ELMo, RoBERTA) due
to the recent acquisition of a more powerful server capable of processing the complete data set
without splitting it into batches. Having more types of embeddings, we can include combinations
of these in the optimization process instead of using a brute-force method. We aim to improve
the performance of our system by further exploring the mentioned steps.
view, Journal of Experimental &amp; Theoretical Artificial Intelligence 0 (2022) 1–
31. URL: https://doi.org/10.1080/0952813X.2022.2153276. doi:10.1080/0952813X.2022.
2153276. arXiv:https://doi.org/10.1080/0952813X.2022.2153276.
[5] M. Á. Álvarez-Carmona, Á. Díaz-Pacheco, R. Aranda, A. Y. Rodríguez-González, L.
BustioMartínez, V. Muñis-Sánchez, A. P. Pastor-López, F. Sánchez-Vega, Overview of rest-mex at
iberlef 2023: Research on sentiment analysis task for mexican tourist texts, Procesamiento
del Lenguaje Natural 71 (2023).
[6] Diccionario de la Lengua Española, 23rd ed., Real Academia Española, 2014. URL: https:
//www.rae.es/.
[7] J. Ortiz-Bejar, V. Salgado, M. Graf, D. Moctezuma, S. Miranda-Jiménez, E. S. Tellez, Ingeotec
at ibereval 2018 task haha:  tc and evomsa to detect and score humor in texts, in:
Proceedings of the Third Workshop on Evaluation of Human Language Technologies for
Iberian Languages (IberEval 2018) co-located with 34th Conference of the Spanish Society
for Natural Language Processing (SEPLN 2018), 2018.
[8] D. Moctezuma, J. Ortiz-Bejar, E. S. Tellez, S. Miranda-Jiménez, M. Graf, Ingeotec solution
for task 4 in tass’18 competition., in: TASS@ SEPLN, 2018, pp. 111–115.
[9] E. S. Tellez, D. Moctezuma, S. Miranda-Jiménez, M. Graf, An automated text categorization
framework based on hyperparameter optimization, Knowledge-Based Systems 149 (2018)
110–123. URL: https://github.com/INGEOTEC/microtc.
[10] T. Mikolov, K. Chen, G. Corrado, J. Dean, L. Sutskever, G. Zweig, word2vec, URL
https://code. google. com/p/word2vec 22 (2013).
[11] T. Mikolov, I. Sutskever, K. Chen, G. S. Corrado, J. Dean, Distributed representations
of words and phrases and their compositionality, in: Advances in neural information
processing systems, 2013, pp. 3111–3119.
[12] J. Devlin, M.-W. Chang, K. Lee, K. Toutanova, Bert: Pre-training of deep bidirectional
transformers for language understanding, arXiv preprint arXiv:1810.04805 (2018).
[13] N. Reimers, I. Gurevych, Making monolingual sentence embeddings multilingual using
knowledge distillation, in: Proceedings of the 2020 Conference on Empirical Methods
in Natural Language Processing, Association for Computational Linguistics, 2020. URL:
https://arxiv.org/abs/2004.09813.
[14] F. A. Acheampong, H. Nunoo-Mensah, W. Chen, Transformer models for text-based
emotion detection: a review of bert-based approaches, Artificial Intelligence Review (2021)
1–41.
[15] T. Chen, C. Guestrin, XGBoost: A scalable tree boosting system, in: Proceedings of the
22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining,
KDD ’16, ACM, New York, NY, USA, 2016, pp. 785–794. URL: http://doi.acm.org/10.1145/
2939672.2939785. doi:10.1145/2939672.2939785.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Álvarez-Carmona</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Aranda</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Arce-Cardenas</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Fajardo-Delgado</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>GuerreroRodríguez</surname>
            ,
            <given-names>A. P.</given-names>
          </string-name>
          <string-name>
            <surname>López-Monroy</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Martínez-Miranda</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Pérez-Espinosa</surname>
            ,
            <given-names>A. Y.</given-names>
          </string-name>
          <article-title>RodríguezGonzález, Overview of rest-mex at iberlef 2021: Recommendation system for text mexican tourism</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <article-title>Álvarez-Carmona, Á</article-title>
          . Díaz-Pacheco,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <string-name>
            <surname>D. FajardoDelgado</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bustio-Martínez</surname>
          </string-name>
          ,
          <article-title>Overview of rest-mex at iberlef 2022: Recommendation system, sentiment analysis and covid semaphore prediction for mexican tourist texts</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>69</volume>
          (
          <year>2022</year>
          )
          <fpage>289</fpage>
          -
          <lpage>299</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Álvarez</surname>
          </string-name>
          <string-name>
            <surname>Carmona</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-Gonzalez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Fajardo-Delgado</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. G.</given-names>
            <surname>Sánchez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Pérez-Espinosa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Martínez-Miranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodríguez</surname>
          </string-name>
          , L. BustioMartínez, Ángel Díaz-Pacheco,
          <article-title>Natural language processing applied to tourism research: A systematic review and future research directions</article-title>
          ,
          <source>Journal of King Saud University - Computer and Information Sciences</source>
          <volume>34</volume>
          (
          <year>2022</year>
          )
          <fpage>10125</fpage>
          -
          <lpage>10144</lpage>
          . URL: https: //www.sciencedirect.com/science/article/pii/S1319157822003615. doi:https://doi.org/ 10.1016/j.jksuci.
          <year>2022</year>
          .
          <volume>10</volume>
          .010.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Diaz-Pacheco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Álvarez</surname>
          </string-name>
          <string-name>
            <surname>Carmona</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. A. C.</given-names>
            <surname>Chávez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Ramírez-Silva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <article-title>Artificial intelligence methods to support the research of destination image in tourism. a systematic re-</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>