<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Pumas PCIC team at REST-MEX 2025:Classification strategies for sentiment analysis</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>León Felipe Dueñas-González</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Cristian Enrique Olvera-Morales</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Helena Gomez-Adorno</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Instituto de Investigaciones en Matemáticas Aplicadas y en Sistemas (IIMAS), Universidad Nacional Autónoma de México</institution>
          ,
          <addr-line>Ciudad de México</addr-line>
          ,
          <country country="MX">México</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Posgrado en Ciencias e Ingeniería de la Computación (PCIC), Universidad Nacional Autónoma de México</institution>
          ,
          <addr-line>Ciudad Universitaria, Coyoacán, 04510</addr-line>
          ,
          <country country="MX">México</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>This paper presents the participation of the Pumas PCIC team in the REST-MEX 2025 shared task, focusing on sentiment analysis and magical town detection using Spanish-language tourist reviews. The study explores traditional models (SVM and XGBoost) and transformer-based architectures (RoBERTa-bne and BETO) to address three classification tasks: sentiment polarity, place type, and town name. Emphasis was placed on mitigating class imbalance through data augmentation techniques such as back-translation and paraphrasing. The final transformer models demonstrated significant improvements in macro F1-score, especially in minority sentiment classes, validating the benefit of phased fine-tuning and contextual enrichment with metadata. The findings highlight the potential of multilingual pre-trained models and domain-specific adaptations in tackling real-world tourism review classification challenges.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Class Imbalance</kwd>
        <kwd>Domain Adaptation</kwd>
        <kwd>Tourism Analytics</kwd>
        <kwd>NLP</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        For Mexico, tourism is a significant economic activity, benefiting both individuals and corporations,
as well as the government. Mexico ofers natural and cultural wealth to all its visitors[
        <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4 ref5">1, 2, 3, 4, 5</xref>
        ].
The “Pueblos Mágicos” program was launched in 2001 during the administration of Vicente Fox. The
ifrst town to be named a “Pueblo Mágico” was Huasca de Ocampo in the state of Hidalgo[ 6]. To be
considered a Magical Town, it must possess symbolic attributes, legends, history, significant events,
everyday life—in short, magic that emanates from each of its sociocultural manifestations, which today
represent a great opportunity for tourism[7].
      </p>
      <p>In this shared task, unlike to other editions [8, 9, 10], we had access to a little over 297,000 reviews
from Tripadvisor, which were delivered in two stages: in the first stage, we received 70% of the data
(just over 208,000), and in the second stage, the remaining 30% (just over 89,000)[11, 12].</p>
      <p>
        Regarding the test dataset, we had the following information:
• Title: The title that the tourist assigned to their opinion. Data type: Text.
• Polarity: The label representing the sentiment polarity of the opinion. Data type: [
        <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4 ref5">1, 2, 3, 4, 5</xref>
        ].
• Region: The state in Mexico where the town is located. Data type: Text.
      </p>
      <p>• Type: The type of place the review refers to. Data type: [Hotel, Restaurant, Attractive].</p>
    </sec>
    <sec id="sec-2">
      <title>2. Summary of Findings from Reviewed Articles</title>
      <p>The first reviewed article corresponds to the work by Morales-Murillo et al. (2023), presented at
RestMex 2023 [13, 10]. This study addresses polarity classification, place type classification, and country of
origin identification in Mexican tourist reviews using Transformer-based models and domain adaptation
techniques.</p>
      <p>The authors employed the RoBERTa-base-bne model, previously pretrained on large Spanish
corpora. Three training strategies were implemented:</p>
      <sec id="sec-2-1">
        <title>1. Direct fine-tuning on the data without robust preprocessing.</title>
        <p>2. Domain adaptation followed by fine-tuning, which achieved the best results.
3. Strategy 2 complemented with data augmentation techniques.</p>
        <p>The dataset was recoded for polarity (from 1–5 to 0–4), and three independent models were
constructed: one for polarity, one for place type, and one for country classification. In the third strategy,
oversampling was applied to balance minority classes in polarity, which improved overall performance.</p>
        <p>Aspect-Based Sentiment Analysis Using CNN</p>
        <p>The second article, published by Pérez et al. (2022) [14], proposes a traditional approach based on
deep convolutional neural networks (CNN) for Aspect-Based Sentiment Analysis (ABSA), using the
dataset from the SemEval-2016 Task 5 challenge.</p>
        <p>The preprocessing stage involved the removal of mentions, URLs, emoticons, and special characters,
followed by lemmatization using spaCy. Vector representations were generated with fastText and
word2vec embeddings.</p>
        <p>Two architectures were designed for aspect detection:
• AB1-CNN (sequential model),
• AB2-CNN (multi-channel model).</p>
      </sec>
      <sec id="sec-2-2">
        <title>And three for sentiment classification:</title>
        <p>• P1-CNN and P3-CNN (sequential models),
• P2-CNN (multi-channel model).</p>
        <p>Key results include an F1-score of 0.6540 for aspect detection using AB1-CNN, and an accuracy of
0.7969 for sentiment classification with P2-CNN. In real-world field tests, combining models achieved
F1-scores up to 0.9333. However, a drop in precision was observed due to polarity class imbalance,
highlighting a significant challenge for future research.</p>
        <p>Transformer-Based Approach for ABSA</p>
        <p>The third article, by Gómez et al. (2024) [15], proposes a more recent model based on Transformer
architectures, specicfially BETO, a BERT model pretrained on large Spanish corpora. The objective
was to enhance ABSA performance through a simple architecture with minimal preprocessing and an
additional linear layer for both aspect detection and sentiment classification.</p>
        <p>Two context selection strategies were evaluated:
• SDR (syntactic dependency relations),
• WW (a 5-word window around the aspect), the latter proving more efective.</p>
        <p>In addition to BETO, other models such as RoBERTa, RoBERTuito, and GPT-2 were compared.
Results show that BETO outperformed others in aspect detection, while RoBERTa combined with WW
performed best in sentiment classification. Despite the simplicity of the architecture, the results reached
state-of-the-art performance in the evaluated tasks.
2.1. Overall Conclusion
The three reviewed studies provide complementary approaches to the problem of sentiment analysis in
Spanish-language texts. The work by Morales-Murillo et al. (2023) demonstrates the efectiveness of
domain adaptation with Transformer models in a real-world tourism context. Meanwhile, the study
by Pérez et al. (2022) shows that traditional CNNs remain useful, especially in scenarios with limited
computational resources. Finally, Gómez et al. (2024) present a substantial improvement by leveraging
the power of Transformer models and modern context selection strategies, setting a new methodological
standard in the field.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Distribution of Variables and Classes</title>
      <p>In order to better understand the structure of the provided dataset, a detailed exploratory analysis
was conducted. This analysis included the graphical representation of each relevant variable, which
allowed us to observe their respective distributions. Additionally, the distribution of the classes and
the composition of the diferent present groups were examined, with the aim of identifying possible
imbalances, relevant patterns, or particular characteristics that could influence the subsequent modeling
process.</p>
      <p>Figure 1 shows the distribution of records according to the type of place referenced in each opinion
within the dataset. It can be observed that the Restaurant category concentrates the highest number of
opinions, with a total of 86,720 records, followed by Attractive with 69,921 records, and finally Hotel,
with 51,410 records. This distribution suggests that the gastronomic experience represents a central
component in the reviews given by tourists, surpassing even opinions about accommodations and
tourist attractions.</p>
      <p>Figure 2 illustrates the distribution of records according to the sentiment polarity expressed in the
reviews. A strong bias towards highly positive evaluations can be observed, with polarity 5 being the
most frequent category, totaling 136,561 records. This is followed in decreasing order by polarity 4
(45,034 records), polarity 3 (15,519 records), while negative evaluations — polarities 1 and 2 — represent
a minimal percentage of the total, with 5,441 and 5,496 records, respectively.</p>
      <p>This behavior demonstrates a clear asymmetry in user perception, suggesting a general tendency to
provide favorable opinions. From a modeling perspective, this highly imbalanced distribution could
afect the performance of classification algorithms, especially in detecting negative reviews. Therefore,
it will be necessary to consider class balancing or adjustment techniques in later stages of the analysis.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Metodology</title>
      <p>Before we started preprocessing the train data, we ran a standard model of SVM and XGBoost for each
category, so we could compare ahead in the processes if it did improve the score. It was decided to
compare with SVM as it is more robust to noise and nonlinear data.It supports multiclass tasks directly
During the development we incorporated two techniques that we will be using in the following tests
that we considered important. The first one consists in combine the columns “Tittle” and “Review”,
by merging this two will be adding more context to the data, enriching the representation of the
text and enhancing the ability of the classification model to capture the meaning and relationship
between keywords. The second is the use of the method Term Frequency Inverse Document Frequency
(TF-IDF) which captures the frequency of each word and penalizes the most common words in all the
corpus. This will aid the model’s ability to diferentiate classes and find relevant information or patterns.
Three variants for each model where tested, the first variants for SVM and XGBoost has a baseline of
max_features = 5000 and ngrams range on 1 and 2. The second we adjusted the previous parameters to
max_features = 10000 and ngrams on 1,3. The last one, we add the class_weight to balanced on SVM
and “scale_pos_weight” so we could penalize the minority classes. The results are shown as follows:</p>
      <p>Preprocessing was a fundamental stage to assure quality and consistency of data. While examine
the data, we detect some texts were written in other languages, so in order to detect this other data
not necessarily fundamental for Spanish analysis, we run a language detector with “Lang detect” and
the result threw that 0.18024% of the data was in other language, most of it were in English and in
Accuracy</p>
      <p>Macro F1
Accuracy</p>
      <p>Macro F1
Accuracy</p>
      <p>Macro F1
polarity 5. Since the majority class for polarity is 5, it was decided not to use this information and
to remove it from the corpus. Then, the text was standardized removing URL, emails and possessive
symbols in English, symbols like ®, ~, ©. and # were also eliminated since doesn’t apport context. For
the hashtags we leave the word that followed it. Some reviews had words that seemed to be part of a
button that displayed more information, this text “. . . Más” wasn’t supposed to be relevant so it was
removed too with some variants presented in most reviews. Character normalization to UTF 8 was also
applied, emoticon substitution for words, use of some slang words were substituted for more reliable
words. For tokenization and lemmatization, spaCy library was used, due to its high eficiency and
accuracy in Spanish, in addition to its ease of integration into processing pipelines. It allows a fast and
suficiently accurate analysis for classification tasks such as those addressed in this work. Based on
Polarity, additional feature extraction for the SVM model was incorporated to complement the TF-IDF
vector-based representation. Emotional and structural attributes, such as the presence of exclamations
and words with negative connotation, were included to capture the subjective intensity of opinions
and improve polarity prediction. Excessive use of exclamation marks were more present in polarity
one and five. Character length did not represent a characteristic of any polarity and words such as
"bad, bad, lousy" were relevant on polarity one and two. Then based on the results of the first runs
we train the model with all the previously mentioned preprocessing on the data and made some other
experiments adding class weights resulting on 0.49% on macro F1 only for polarity. Next, trying t to
solve the imbalance in classes , it was added the “RandomOverSampler” with a seed of 42 which did
not improve the previous score with 3% less. For the last adjustment on both models, for Polarity the
input embedding was trained using a representation that combined the lemmatized text vectorized
with TF-IDF, the variable Region encoded with One Hot Encoding and the emotional and structural
features (negative words, text length and exclamation marks). This combination allowed us to capture
both semantic features and subjective intensity and geographical context. Type and Town on the other
hand were only trained with region and the TF-IDF vectors, this configuration proved to be efective
for classifying the categories of place type and magic town without adding the additional subjective
intensity characteristics, which were less relevant for these tasks. Additionally for Town we detected
that 20.03% of the reviews contained the value of “Town” so we develop an additional function to classify
it if this was the case. The results for the final 2 models are as follows:</p>
      <sec id="sec-4-1">
        <title>Model</title>
        <p>SVM
XGBoost
SVM
XGBoost
SVM
XGBoost</p>
      </sec>
      <sec id="sec-4-2">
        <title>Task</title>
        <p>Polarity
Polarity</p>
        <p>Town
Town
Type
Type</p>
        <p>The previous models showed good results for some classes, but they presented some limitations
capturing the context semantics and long-range relationships within text. In comparison, Transformer
models leverage attentional mechanisms that allow modeling complex and contextual dependencies more
efectively. This is especially relevant in a domain with informal language and colloquial expressions
such as Spanish tourist reviews, where the meaning of a sentence may depend on its surrounding
context. This facilitates knowledge transfer by being pre-trained on large corpora, which improves
generalization to new examples. Therefore, we chose to integrate Transformers models to take advantage
of these in the classification task, obtaining notable improvements in precision and recall with respect
to traditional methods. In order to solve the problem of imbalanced data in polarity classification,
augmented data techniques were implemented for minority classes such as one, two and three. The
dataset was split into 90% train and 10% validation. We applied to train data the back translation
technique using Helsinki-NLP/opus-mt to automatically translate to English and backwards to Spanish,
also T5 was used to paraphrasing. This generated 30,431 new examples for training which were cleaned
and preprocessed after the previous process generated some noise. The pre trained models used in this
work were RoBERTa-bne and BETO both in Spanish, which help us to capture linguistic particularities.
To reinforce the problem address for polarity, CrossEntropyLoss function was used to compensate the
inter-class imbalance. To calculate the weights, the formula  =  where N is the total number
of examples in the corpus and  is the number of examples for class c. Tokenizing the combination
of title and review, the model input was also enriched by adding additional contextual information
from the region and type columns. We applied two runs trained using the Hugging Face Trainer class
with a carefully selected configuration. A learning rate of 2e-5 and a batch size of 16, applying a
weight_decay of 0.01 to regularize the model. To facilitate the selection of the best model, the parameter
load_best_model_at_end=True was set and f1_macro was used as the main metric were used for both
training and validation with the only diference that the first run was trained with the train set plus the
augmented data during 6 epoch and the second for four epochs only with train data and a fine tuning
for 3 epoch with only the augmented data.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Automatic Classification System for Tourist Reviews</title>
      <p>An automatic classification system was developed to identify both the type of site reviewed and the
Pueblo Mágico associated with each opinion. The reviews could refer to hotels, restaurants, or attractions,
and the dataset required an initial preprocessing phase. This phase included the cleaning of special
characters, normalization of feminine ordinals, and mapping of the target variables to facilitate their
use in machine learning models.</p>
      <p>For the task of site type classification, the base model used was RoBERTa-large-bne, a large-scale
version of the RoBERTa model adapted to Spanish. This model was pre-trained on a 570 GB corpus of
clean, deduplicated text collected by the Spanish National Library between 2009 and 2019. A two-stage
domain adaptation process was applied. In the first stage, the model was adjusted using the training
dataset specific to the task, and in the second stage, an external corpus was used to further align the
model with the tourism domain. Once the domain adaptation was completed, fine-tuning was performed
for a three-class classification task, using five training epochs, a learning rate of 2e-5, and a batch size of
64 samples per step. The model achieved an average F1 score of 0.982 on the validation set and 0.976
on the test set.</p>
      <p>For the identification of the associated Pueblo Mágico, a two-stage strategy was implemented. The first
stage involved detecting direct mentions of the town’s name within the text, including name variants
and counting occurrences to determine the most likely match. This stage achieved an F1 score of 0.9753.
For cases without a direct mention, a second strategy was applied, based on feature extraction using
ID-TF over the text, title, and region, followed by classification with a multiclass SVM model capable of
distinguishing between 40 diferent Pueblos Mágicos.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Results</title>
    </sec>
    <sec id="sec-7">
      <title>7. Discussion</title>
      <p>As a result, we obtained an improvement in F1 macro and in classes 2 and 3, which validates the phased
approach as an alternative strategy in the face of imbalance [Table 5]
Transformers demonstrated significant advantages in modeling complex contextual relationships and
outperformed traditional models in accuracy and recall, consolidating their choice as the final solution
in this challenge.</p>
    </sec>
    <sec id="sec-8">
      <title>8. Conclusions</title>
      <p>Throughout the development of solutions for this contest, we developed an SVM classifier due to its
simplicity and easy interpretation. However, the model did not achieve the expected results,
especially for the class imbalance, linguistic variability, and the final result for town prediction. Despite
hyperparameter tuning and preprocessing techniques, performance remained below expectations for
metrics such as F1 score and recall for minority classes. For “town” classification task, we decided to
use an SVM classifier due to the large number of classes and the constraints of time and computational
resources. Although it managed to generate acceptable predictions on the tests, its final results did
not exceed performance expectations on the oficial contest test set. Although our hybrid approach on
Transformers and SVM generated a solution, the results obtained with SVM on the "town" task suggest
that a transformer model would have represented a more robust and coherent option with the rest of
our architecture, possibly achieving more consistent and competitive performance.</p>
    </sec>
    <sec id="sec-9">
      <title>Declaration on Generative AI</title>
      <p>We declare that the present manuscript has been written entirely by the authors and that no generative
artificial intelligence tools were used in its preparation, drafting, or editing.
[6] A. M. F. Poncela, Una revisión del programa pueblos mágicos, CULTUR: Revista de Cultura e</p>
      <p>Turismo 10 (2016) 3–34.
[7] M. L. D. Torres, J. C. S. Salinas, J. M. Arias, H. V. Lapaz, J. de Dios Romero Palop, Big data y
turismo en méxico: Pueblos mágicos, 2016. URL: https://datatur.sectur.gob.mx/Documentos%
20compartidos/2016_5.pdf.
[8] M. Á. Álvarez-Carmona, R. Aranda, S. Arce-Cárdenas, D. Fajardo-Delgado, R. Guerrero-Rodríguez,
A. P. López-Monroy, J. Martínez-Miranda, H. Pérez-Espinosa, A. Rodríguez-González, Overview
of rest-mex at iberlef 2021: Recommendation system for text mexican tourism, Procesamiento del
Lenguaje Natural 67 (2021). doi:https://doi.org/10.26342/2021-67-14.
[9] M. Á. Álvarez-Carmona, Á. Díaz-Pacheco, R. Aranda, A. Y. Rodríguez-González, D. Fajardo-Delgado,
R. Guerrero-Rodríguez, L. Bustio-Martínez, Overview of rest-mex at iberlef 2022:
Recommendation system, sentiment analysis and covid semaphore prediction for mexican tourist texts,
Procesamiento del Lenguaje Natural 69 (2022) 289–299.
[10] M. Á. Álvarez-Carmona, Á. Díaz-Pacheco, R. Aranda, A. Y. Rodríguez-González, V. Muñiz-Sánchez,
A. P. López-Monroy, F. Sánchez-Vega, L. Bustio-Martínez, Overview of rest-mex at iberlef 2023:
Research on sentiment analysis task for mexican tourist texts, Procesamiento del Lenguaje Natural
71 (2023) 425–436.
[11] M. Á. Álvarez-Carmona, Á. Díaz-Pacheco, R. Aranda, A. Y. Rodríguez-González, L. Bustio-Martínez,
V. Herrera-Semenets, Overview of rest-mex at iberlef 2025: Researching sentiment evaluation in
text for mexican magical towns, volume 75, 2025.
[12] J. Á. González-Barba, L. Chiruzzo, S. M. Jiménez-Zafra, Overview of IberLEF 2025: Natural
Language Processing Challenges for Spanish and other Iberian Languages, in: Proceedings of the
Iberian Languages Evaluation Forum (IberLEF 2025), co-located with the 41st Conference of the
Spanish Society for Natural Language Processing (SEPLN 2025), CEUR-WS. org, 2025.
[13] V. G. Morales-Murillo, H. Gómez-Adorno, D. Pinto, I. A. Cortés-Miranda, P. Delice, Lke-iimas
team at rest-mex 2023: Sentiment analysis on mexican tourism reviews using transformer-based
domain adaptation (2023).
[14] J. Pérez, M. Sánchez, L. Torres, Deep learning approach for aspect-based sentiment analysis of
restaurants reviews in spanish, Journal of Computational Linguistics Research 12 (2022) 45–58.
[15] C. Gómez, L. Ramírez, S. Ortega, Enhancing spanish aspect-based sentiment analysis through
deep learning approach, Natural Language Processing Advances 15 (2024) 22–39.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodriguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Álvarez</surname>
          </string-name>
          <string-name>
            <surname>Carmona</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. P.</given-names>
            <surname>López-Monroy</surname>
          </string-name>
          ,
          <article-title>Studying online travel reviews related to tourist attractions using nlp methods: the case of guanajuato, mexico</article-title>
          ,
          <source>Current Issues in Tourism</source>
          <volume>26</volume>
          (
          <year>2023</year>
          )
          <fpage>289</fpage>
          -
          <lpage>304</lpage>
          . URL: https://doi.org/10.1080/13683500.
          <year>2021</year>
          .
          <volume>2007227</volume>
          . doi:
          <volume>10</volume>
          .1080/13683500.
          <year>2021</year>
          .
          <volume>2007227</volume>
          . arXiv:https://doi.org/10.1080/13683500.
          <year>2021</year>
          .
          <volume>2007227</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Álvarez-Carmona</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          , et al.,
          <article-title>Big data analytics of online news to explore destination image using a comprehensive deep-learning approach: a case from mexico</article-title>
          ,
          <source>Information Technology &amp; Tourism</source>
          <volume>26</volume>
          (
          <year>2024</year>
          )
          <fpage>147</fpage>
          -
          <lpage>182</lpage>
          . URL: https://doi.org/10.1007/ s40558-023-00278-5. doi:
          <volume>10</volume>
          .1007/s40558-023-00278-5.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Á.</given-names>
            <surname>Díaz-Pacheco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Álvarez-Carmona</surname>
            ,
            <given-names>A. Y.</given-names>
          </string-name>
          <string-name>
            <surname>Rodríguez-González</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Aranda</surname>
          </string-name>
          ,
          <article-title>A comprehensive deep learning approach for topic discovering and sentiment analysis of textual information in tourism</article-title>
          ,
          <source>Journal of King Saud University-Computer and Information Sciences</source>
          <volume>35</volume>
          (
          <year>2023</year>
          )
          <fpage>101746</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Diaz-Pacheco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Álvarez-Carmona</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Carlos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <article-title>Measuring the diference between pictures from controlled and uncontrolled sources to promote a destination. a deep learning approach (</article-title>
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>C. M. A.</given-names>
            <surname>Molinar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. G.</given-names>
            <surname>Sánchez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I. M.</given-names>
            <surname>Carrillo</surname>
          </string-name>
          ,
          <article-title>Influencia de la inseguridad en la competitividad turística: el caso de méxico</article-title>
          ,
          <source>RITUR-Revista Iberoamericana de Turismo</source>
          <volume>15</volume>
          (
          <year>2025</year>
          )
          <fpage>34</fpage>
          -
          <lpage>49</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>