<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>BERT-Based Models for Joint Sentiment, Type, and Location Classification of Spanish Tourist Reviews</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Darían Santiago Llanes Guilarte</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Vitali Herrera-Semenets</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lázaro Bustio-Martínez</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Miguel Ángel Álvarez-Carmona</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Advanced Technologies Application Center (CENATAV)</institution>
          ,
          <addr-line>La Habana</addr-line>
          ,
          <country country="CU">Cuba</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Centro de Investigación en Matemáticas</institution>
          ,
          <addr-line>Monterrey</addr-line>
          ,
          <country country="MX">México</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Iberoamerican University</institution>
          ,
          <addr-line>Ciudad de México</addr-line>
          ,
          <country country="MX">México</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>This paper presents two approaches to jointly classify sentiment polarity, tourist place type, and the corresponding Magical Town from Spanish-language reviews of Mexican tourist destinations. Our methods leverage a multi-task neural network based on the TabularisAI multilingual sentiment model (768-dim BERT-base architecture) and a pre-trained BERT model adapted for Spanish. Unlike previous approaches that relied on a unified label space or separate models for each task, we adopt a multi-head architecture that simultaneously optimizes for all three tasks using task-specific classification heads. The system incorporates class balancing through weighted loss functions and advanced preprocessing with spaCy. Evaluation on the oficial Rest -Mex 2025 dataset demonstrates competitive performance, achieving promising results across tasks, while maintaining eficiency and modularity.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Research on sentiment analysis in the Spanish tourism domain has significantly advanced through the
adoption of pre-trained transformer-based models such as BERT [
        <xref ref-type="bibr" rid="ref6 ref7 ref8">6, 7, 8</xref>
        ]. Vásquez employed BETO
combined with TF-IDF weighting over TripAdvisor reviews and achieved first place in Rest-Mex 2021
using a monolingual architecture optimized for Spanish [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. Jurado-Buch et al. proposed a unified
model based on BETO to jointly predict polarity, type, and location in one pass, placing within the top
eight systems at Rest-Mex 2023 [10]. Moreover, Álvarez-Carmona et al. provided a comprehensive
overview of Rest-Mex 2023, ofering insight into dataset characteristics and evaluation protocols [11].
      </p>
      <p>Domain-adapted language models have also been explored. Campos and Viñaña-Ludeña trained a
BERT model specifically for tourism (Spanish-Tourism-BERT) on social media data to extract
locationbased entities and sentiment components [12]. Bouabdallaoui et al. compared BERT fine-tuning
against hybrid models combining sentence embeddings and LSTM networks in Moroccan tourism data,
concluding that BERT fine-tuning yields superior accuracy [ 13]. In terms of multi-task learning, Zhang
et al. implemented a hard-sharing architecture combining a shared BERT encoder with task-specific
layers, showing statistical improvements in multi-output sentiment scenarios [14].</p>
      <p>The reviewed literature reveals three clear trends: (1) fine-tuning monolingual BERT models such as
BETO is highly efective in Spanish tasks; (2) domain-specific models show promise for tourism-related
content; and (3) multi-task architectures provide performance benefits for correlated classification tasks.
However, most works address only one or two subtasks, and none explore simultaneous prediction of
polarity, place type, and town name, as required by Rest-Mex 2025.</p>
      <sec id="sec-2-1">
        <title>2.1. Remarks</title>
        <p>
          The reviewed studies expose several critical gaps that justify the contributions of the present work:
• Most prior works (e.g., [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ], [12]) focus exclusively on a single task such as polarity or aspect
identification. Our proposed architecture performs simultaneous classification of sentiment
polarity, place type, and geographical location, addressing the complete Rest-Mex 2025 task.
• Jurado-Buch et al. addressed multi-label output via a unified label space (45-class combinations),
but their model lacks separate heads or task-specific loss components. Our solution employs
a shared encoder with specialized heads per task and customized loss weighting to enhance
adaptability and generalization.
• While Spanish-Tourism-BERT represents a step toward domain-specific pre-training, it was
tested only on social media and not on structured tourism reviews. Our models include both
multilingual (TabularisAI) and monolingual (BETO) BERT variants adapted through preprocessing,
task weighting, and stratified training.
• Although Zhang et al. introduced a valid multitask architecture, they did not apply it to tourism
datasets or large-scale multilingual corpora. Our study extends both dataset size and architectural
robustness via gradient accumulation, warm-up phases, and diferential learning rates.
After revising the related work, the contributions of the presented work include:
1. Joint evaluation of polarity, place type, and location via multi-head classification.
2. Comparative analysis of monolingual and multilingual BERT-based architectures adapted to
tourism.
3. Implementation of advanced preprocessing and weighted loss strategies for handling extreme
class imbalance.
        </p>
        <p>This work fills evident gaps in the literature and ofers an integrated framework suitable for large-scale,
multidimensional sentiment analysis in Spanish-language tourist reviews.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Dataset</title>
      <p>
        The dataset provided by the Rest-Mex 2025 organizers consists of 208,051 Spanish-language reviews [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
Each review includes a title and a body, which we concatenate into a single text field. The dataset is
annotated with:
• Sentiment polarity on a 5-point scale (1 = very negative, 5 = very positive)
• Type of tourist place: Hotel, Restaurant, or Attractive
• Name of the Magical Town (from a set of 60 distinct towns)
• Region to which the town belongs
      </p>
      <p>A detailed exploratory data analysis (EDA) was conducted to understand the distribution of the data.
Key findings include:
• Polarity distribution: The data is heavily skewed toward positive reviews, with over 65% labeled
as polarity 5 (see Figure 1). Only about 2.6% are rated as polarity 1.
• Type distribution: The most common category is Restaurant (86,720 instances), followed by</p>
      <p>Attractive (69,921), and Hotel (51,410) (see Figure 2).
• Top towns: Tulum (45,345 reviews), Isla Mujeres (29,826), and San Cristóbal de las Casas (13,060)
are the towns with the most reviews (see Figure 3).
• Top regions: Quintana Roo leads with 85,993 reviews, followed by Chiapas (23,532), and Estado
de México (19,439).
• Text length: The average length of concatenated Title + Review texts is 65 words, with a
maximum of 1,487 and a minimum of 2 words (see Figure 4).
• Missing values: Only the Title field had missing values (2 instances), which were handled by
exclusion.</p>
      <p>These findings reveal significant class imbalance, particularly in the polarity label, justifying the use
of class-weighted losses in training.</p>
      <p>We use 80% of the dataset for training and 20% for validation, following a stratified sampling strategy
to preserve label distributions.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Methodology</title>
      <p>This section describes our two proposals evaluated on the oficial Rest-Mex 2025 dataset.</p>
      <sec id="sec-4-1">
        <title>4.1. Hammer Squat_1_Run</title>
        <p>This approach leverages a multilingual sentiment analysis model integrated with optimized text
preprocessing and multi-task learning. Specifically, our solution incorporates three core components:
advanced linguistic normalization, a transformer-based architecture with task-specific heads, and
adaptive training strategies for class imbalance mitigation.</p>
        <sec id="sec-4-1-1">
          <title>4.1.1. Data Preprocessing</title>
          <p>Text normalization techniques were systematically applied to raw inputs to ensure uniform
formatting. Subsequently, the title and review fields were concatenated to form unified text representations.</p>
          <p>Since reviews constituted our primary focus, encoding correction was performed using ftfy
alongside emoji normalization via emoji.demojize. Additionally, non-informative elements including
URLs, user mentions, and domain-specific stopwords were removed while preserving diacritics and
numerical values. Character repetition patterns were concurrently reduced to single instances. Finally,
language detection excluded non-Spanish texts, with complementary length-based filtering retaining
texts exceeding 25 characters and 5 meaningful tokens.</p>
        </sec>
        <sec id="sec-4-1-2">
          <title>4.1.2. Model Architecture</title>
          <p>The architecture utilized tabularisai/multilingual-sentiment-analysis, a BERT-base model
pretrained for multilingual sentiment tasks [15]. Contextual representations from this shared encoder
subsequently fed three specialized classification heads: a 5-dimensional output for polarity prediction
(1-5 star scale), a 3-dimensional output for establishment type classification (Hotel, Restaurant, or
Attractive), and a 40-dimensional output for location identification (Magical Towns). To address dataset
imbalances, class weights based on inverse frequency were applied to balance the loss functions.</p>
        </sec>
        <sec id="sec-4-1-3">
          <title>4.1.3. Training Configuration</title>
          <p>AdamW optimization employed diferentiated learning rates ( 1 × 10− 3 for classification heads versus
2 × 10− 5 for the base encoder). Furthermore, loss components were weighted by task complexity:
30% polarity, 30% establishment type, and 40% location. Gradient accumulation (over 2 steps) enabled
efective batch sizes of 16, while gradient clipping (max norm 1.0) stabilized convergence. The learning
rate schedule combined a 10% warm-up phase with progressive encoder unfreezing after epoch 1.
Ultimately, mixed-precision training accelerated computation throughout 4 epochs.</p>
        </sec>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Hammer Squat_2_Run</title>
        <p>In this section is presented our second approach (_2_). Our second proposal
consists of three key components: data preprocessing, model design, and training configuration. Each
step plays a critical role in building an efective and robust multi-task learning system.</p>
        <sec id="sec-4-2-1">
          <title>4.2.1. Data Preprocessing</title>
          <p>Preprocessing is essential to convert noisy, unstructured text into a clean format that machine learning
models can interpret efectively. First, we concatenate the Title and Review fields to form a unified
text input for the model. This ensures the model has access to all user-generated content associated
with a review.</p>
          <p>Next, we apply linguistic preprocessing using spaCy’s es_core_news_sm model. This includes
lemmatization (reducing words to their base forms), stopword removal (to eliminate uninformative
words), and punctuation filtering. These steps reduce noise and vocabulary size, improving the model’s
generalization.</p>
          <p>Finally, we encode the categorical labels (polarity, type, and town) using label encoding,
transforming them into numerical values compatible with neural networks.</p>
        </sec>
        <sec id="sec-4-2-2">
          <title>4.2.2. Model Architecture</title>
          <p>We adopt a multi-task learning approach built on top of the pre-trained BERT model for Spanish:
bert-base-spanish-wwm-cased [16]. Multi-task learning enables the model to learn shared
representations that benefit multiple related tasks in this case, predicting polarity, place type, and town.</p>
          <p>The architecture consists of:
• A shared BERT encoder that processes the input text into contextual embeddings.
• Three separate classification heads fully connected layers for each prediction task:
– Polarity: 5 output neurons (for classes 1 to 5)
– Type: 3 output neurons (hotel, restaurant, attraction)
– Town: 60 output neurons (corresponding to each Magical Town)</p>
          <p>This design allows each head to specialize while benefiting from shared information learned across
tasks. Each head is optimized independently via task-specific cross-entropy loss. Also, class imbalance
is mitigated by computing class weights and integrating them into each loss function. The multi-head
architecture improves model eficiency by avoiding the need to train separate models.</p>
        </sec>
        <sec id="sec-4-2-3">
          <title>4.2.3. Training Configuration</title>
          <p>To train the model, we use the AdamW optimizer with a learning rate of 2e-5, which is well-suited
for nfie-tuning transformers. We apply the ReduceLROnPlateau scheduler to automatically reduce the
learning rate when validation performance plateaus, ensuring stable convergence.</p>
          <p>A batch size of 16 is chosen to balance performance and memory usage, and we train for 15 epochs.
We also calculate class weights based on the label distribution and use them in the cross-entropy loss
function for each task. This addresses class imbalance and encourages the model to pay more attention
to minority classes.</p>
          <p>During training, the model is evaluated on a validation split to track performance. The best-performing
model checkpoint is saved and later used for evaluation and inference.</p>
          <p>Together, these methodological components ensure that the model is well-prepared to handle the
complexities of real-world multilingual, multi-label classification in the tourism domain.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
      <p>We assess the system’s performance Macro F1-score for each task. As shown in Figure 1, the results
obtained by the Hammer Squat team in the REST-MEX competition, represented by the solutions Hammer
Squat_1_Run (20th place) and Hammer Squat_2_Run (41st place), show a considerable performance
gap compared to the top three entries: UDENAR_1, Axolotux_E_T3, and Axolotux_E3.</p>
      <p>While both Hammer Squat submissions achieved reasonable scores, there are clearly critical areas
for improvement, particularly in the Macro F1 (Polarity) and Macro F1 (Town) metrics, where the most
notable diferences can be observed. For instance, in the polarity task, Hammer Squat_1_Run scored
0.579, compared to 0.644 by the top-ranking model. The gap is even wider for Hammer Squat_2_Run,
which achieved only 0.473 in this metric. Regarding the town classification task (Macro F1 (Town)),
the Hammer Squat models again lag behind, with scores of 0.580 and 0.441, while the leading model
reached 0.692.</p>
      <p>Notably, in the Macro F1 (Type) metric, which evaluates entity type classification, the Hammer Squat
models performed more competitively (0.970 and 0.944) compared to the winner (0.987), indicating that
the team’s approach has strengths in certain specific tasks. Overall, while the Hammer Squat solutions
did not reach the top ranks, the results show promising potential, particularly if improvements are made
in components related to sentiment detection and geographic localization. These key areas could benefit
from further refinement, such as implementing more advanced linguistic preprocessing techniques or
leveraging models specifically tuned for sentiment analysis and textual geolocation.</p>
      <p>Comprehensive details regarding the overall results, including information about the participating
teams, can be found in [17].</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>This paper introduce two models for sentiment analysis of Spanish-language reviews in the tourism
domain. Our system jointly predicts sentiment polarity, place type, and location with promising results
using shared representations and task-specific outputs. The approaches are eficient and adaptable,
achieving high performance with modest hardware requirements. Future directions include integrating
attention-based fusion and exploring multilingual generalization.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgements</title>
      <p>The authors gratefully acknowledge the support provided by the Mexican Academy of Tourism
Research (AMIT) for the project “Balancing Tourism Text Data with Artificial Intelligence for Sentiment
Analysis: A Specialized Language Model Approach” funded through the Research Projects 2024 call.
Additionally, this work was also supported by the project “Text Generation for Data Balancing in
Sentiment Classification: Application to Tourism Data” under the CICIMPI 2024 call of the Centro de
Investigación en Matemáticas (CIMAT).</p>
    </sec>
    <sec id="sec-8">
      <title>Declaration on Generative AI</title>
      <p>We declare that the present manuscript has been written entirely by the authors and that no generative
artificial intelligence tools were used in its preparation, drafting, or editing.
spanish reviews from tripadvisor, in: IberLEF 2021, CEUR Workshop Proceedings, 2021. 1st place
Rest-Mex 2021.
[10] J. D. Jurado-Buch, E. S. Minayo-Díaz, J. A. Tello, et al., Juan david jurado-buch and ever sebastian
minayo-díaz and jhony alexander tello and others, in: Rest-Mex 2023, CEUR Workshop Proceedings,
2023. Top-8 Rest-Mex 2023.
[11] M. Ángel Álvarez Carmona, Ángel Díaz-Pacheco, R. Aranda, L. Bustio-Martínez, Overview of
rest-mex at iberlef 2023: Research on sentiment analysis task for mexican tourist texts, CEUR
Workshop Proceedings (2023).
[12] M. S. V. naña Ludeña, L. M. de Campos, Discovering a tourism destination with social media data:</p>
      <p>Bert-based sentiment analysis, Journal of Hospitality and Tourism Technology (2021).
[13] I. Bouabdallaoui, F. Guerouate, S. Bouhaddour, C. Saadi, M. Sbihi, Sentiment analysis in tourism:</p>
      <p>Fine-tuning bert or sentence embeddings concatenation?, arXiv preprint arXiv:2302.04519 (2023).
[14] K. Y. Zhang, J. Zhang, Y. Mao, Multi-task learning for sentiment analysis with hard-sharing and
task recognition mechanism, Information 12 (2021) 207.
[15] tabularisai.multilingual-sentiment analysis, Distilbert-based multilingual sentiment classification
model, Available: https://huggingface.co/tabularisai/multilingual-sentiment-analysis (2024).
[16] J. Cañete, G. Chaperon, R. Fuentes, J.-H. Ho, H. Kang, J. Pérez, Spanish pre-trained bert model and
evaluation data, in: PML4DC at ICLR 2020, 2020.
[17] Rest-Mex, Results, Available: https://sites.google.com/cimat.mx/rest-mex-2025/results (2025).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Álvarez-Carmona</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Aranda</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Arce-Cárdenas</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Fajardo-Delgado</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Guerrero-Rodríguez</surname>
            ,
            <given-names>A. P.</given-names>
          </string-name>
          <string-name>
            <surname>López-Monroy</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Martínez-Miranda</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Pérez-Espinosa</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <article-title>Overview of rest-mex at iberlef 2021: Recommendation system for text mexican tourism</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>67</volume>
          (
          <year>2021</year>
          ). doi:https://doi.org/https://doi.org/10.26342/2021-67-14.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <article-title>Álvarez-Carmona, Á</article-title>
          . Díaz-Pacheco,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Fajardo-Delgado</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bustio-Martínez</surname>
          </string-name>
          ,
          <article-title>Overview of rest-mex at iberlef 2022: Recommendation system, sentiment analysis and covid semaphore prediction for mexican tourist texts</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>69</volume>
          (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <article-title>Álvarez-Carmona, Á</article-title>
          . Díaz-Pacheco,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bustio-Martínez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Muñis-Sánchez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. P.</given-names>
            <surname>Pastor-López</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Sánchez-Vega</surname>
          </string-name>
          ,
          <article-title>Overview of rest-mex at iberlef 2023: Research on sentiment analysis task for mexican tourist texts</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>71</volume>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Rest-Mex</surname>
          </string-name>
          , Rest-mex
          <year>2025</year>
          :
          <article-title>Researching sentiment evaluation in text for mexican magical towns</article-title>
          , Available: https://sites.google.com/cimat.mx/rest-mex-2025/home (
          <year>2025</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>González-Barba</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Chiruzzo</surname>
            ,
            <given-names>S. M.</given-names>
          </string-name>
          <string-name>
            <surname>Jiménez-Zafra</surname>
          </string-name>
          ,
          <article-title>Overview of IberLEF 2025: Natural Language Processing Challenges for Spanish and other Iberian Languages, in: Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2025), co-located with the 41st Conference of the Spanish Society for Natural Language Processing (SEPLN 2025), CEUR-WS</article-title>
          . org,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Álvarez-Carmona</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-Gonzalez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Fajardo-Delgado</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. G.</given-names>
            <surname>Sánchez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Pérez-Espinosa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Martínez-Miranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bustio-Martínez</surname>
          </string-name>
          ,
          <article-title>Ángel DíazPacheco, Natural language processing applied to tourism research: A systematic review and future research directions</article-title>
          ,
          <source>Journal of King Saud University - Computer and Information Sciences</source>
          <volume>34</volume>
          (
          <year>2022</year>
          )
          <fpage>10125</fpage>
          -
          <lpage>10144</lpage>
          . URL: https://www.sciencedirect.com/science/article/pii/S1319157822003615. doi:https://doi.org/https://doi.org/10.1016/j.jksuci.
          <year>2022</year>
          .
          <volume>10</volume>
          .010.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Á.</given-names>
            <surname>Díaz-Pacheco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Álvarez-Carmona</surname>
            ,
            <given-names>A. Y.</given-names>
          </string-name>
          <string-name>
            <surname>Rodríguez-GonzÁlez</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Aranda</surname>
          </string-name>
          ,
          <article-title>A comprehensive deep learning approach for topic discovering and sentiment analysis of textual information in tourism</article-title>
          ,
          <source>Journal of King Saud University - Computer and Information Sciences</source>
          <volume>35</volume>
          (
          <year>2023</year>
          )
          <article-title>101746</article-title>
          . URL: http://dx.doi.org/10.1016/j.jksuci.
          <year>2023</year>
          .
          <volume>101746</volume>
          . doi:https://doi.org/ 10.1016/j.jksuci.
          <year>2023</year>
          .
          <volume>101746</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Álvarez-Carmona</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          , et al.,
          <article-title>Big data analytics of online news to explore destination image using a comprehensive deep-learning approach: a case from mexico</article-title>
          ,
          <source>Information Technology &amp; Tourism</source>
          <volume>26</volume>
          (
          <year>2024</year>
          )
          <fpage>147</fpage>
          -
          <lpage>182</lpage>
          . URL: https://doi.org/10.1007/ s40558-023-00278-5. doi:https://doi.org/10.1007/s40558-023-00278-5.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J.</given-names>
            <surname>Vásquez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Gómez-Adorno</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Bel-Enguix</surname>
          </string-name>
          ,
          <article-title>Bert-based approach for sentiment analysis of</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>