<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Journal of King Saud University</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.1080/13683500.2021.2007227</article-id>
      <title-group>
        <article-title>A Parallel NLP Pipeline with NER-Enhanced Hierarchical Classification: Sentiment Analysis for Mexican Magical Towns</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Gabriel Santiago Robles-Robles</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jehu Jonathan Ramirez-Ramirez</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Angel David Durazo-Bartolini</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gael Balderrama-Dominguez</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Luis Hiram Hernandez-Gutierrez</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Victor Hugo Ramirez-Rios</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Juan Adan Nava-Banda</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Gustavo Gutierrez-Navarro</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Juan Daniel Garcia-Ruiz</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mario Alejandro Castro-Lerma</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jesus Antonio Flores-Briones</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Manuel Ivan Melendez-Rivera</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Christian Alexis Flores-Alvarez</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jose Angel Morales-Nuñez</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mauricio Toledo-Acosta</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Universidad de Sonora, Departamento de Matemáticas, Blvd. Luis Encinas y Rosales, Col. Centro</institution>
          ,
          <addr-line>Hermosillo, Sonora, 83000</addr-line>
          ,
          <country country="MX">Mexico</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2014</year>
      </pub-date>
      <volume>10</volume>
      <fpage>10125</fpage>
      <lpage>10144</lpage>
      <abstract>
        <p>We present a parallel NLP pipeline for sentiment analysis and multi-label classicfiation of Spanish tourism reviews from Mexican Magical Towns, developed for the Rest-Mex 2025 shared task. Our architecture combines three specialized models: (1) a fine-tuned Qwen3-0.6B transformer for 5-class sentiment prediction, (2) a TF-IDF logistic regression classifier for destination type categorization, and (3) a NER-enhanced hierarchical model for town identification that integrates named entity recognition with BERT embeddings. The system achieved 69.57th percentile overall in the competition, with the town classifier excelling in location-specific performance (Macro F1: 0.6006, 75.36th percentile). Macro F1 scores of 0.4753 for sentiment analysis and 0.9423 for destination type classification demonstrate the efectiveness of our modular approach. Key contributions include handling extreme class imbalance (e.g., 62:1 in the town labels distribution) through hierarchical classification. Results demonstrate that hybrid architectures combining transformers, traditional Machine Learning, and knowledge-enhanced components outperform monolithic approaches for tourism NLP tasks, though challenges remain in fine-grained sentiment analysis. Our modular design ofers computational eficiency while maintaining a certain degree of interpretability.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Tourism Sentiment Analysis</kwd>
        <kwd>Modular NLP Pipelines</kwd>
        <kwd>Named Entity Recognition</kwd>
        <kwd>Large Language Models</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Tourism remains a critical driver of global economic growth, contributing significantly to employment
and GDP worldwide. In the past decade, the industry has increasingly shifted its operations online,
with digital platforms playing a pivotal role in shaping travel decisions [1, 2, 3, 4]. In Mexico, tourism
continues to be a cornerstone of the economy, accounting for approximately 8.6% of GDP [5], and
generating over 4.8 million direct jobs as of the second quarter of 2024 [5]. Despite the severe disruptions
caused by the COVID-19 pandemic, the national sector has shown resilience, with recovery trends
highlighting the growing reliance on data-driven strategies to adapt to evolving traveler behaviors
[5, 6, 7].</p>
      <p>In this context, Artificial Intelligence (AI)—particularly Natural Language Processing (NLP)—has
emerged as a crucial tool for extracting insights from tourists’ opinions [8, 9]. Social media platforms,
review websites, and other digital channels generate vast amounts of unstructured feedback, reflecting
subjective experiences, sentiments, and preferences [10, 11, 12]. For instance, recent studies demonstrate
that a vast majority of travelers rely on online reviews when planning trips [1], while businesses and
policymakers increasingly leverage this data to enhance services, optimize marketing, and design
targeted policies [7].</p>
      <p>This paper addresses sentiment analysis and multi-label classification of Spanish-language tourism
reviews from Mexican Magical Towns as part of the Rest-Mex 2025: Researching Sentiment Evaluation
in Text for Mexican Magical Towns at IberLef 2025. Building on advances in NLP, we propose a parallel
processing architecture consisting of three specialized models: a fine-tuned Qwen3-0.6B transformer for
sentiment polarity classification (1-5 scale), a TF-IDF-based logistic regression classifier for destination
type prediction (hotel, restaurant, attraction), and a hybrid two-stage model for town identification
that combines named entity recognition with hierarchical BERT-based classification across 40 Mexican
Magical Towns.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>Sentiment analysis has evolved from traditional rule-based approaches to more sophisticated deep
learning architectures, with tourism-specific applications gaining considerable attention in recent
years [13]. The field encompasses three primary methodological categories: knowledge-based systems,
machine learning approaches, and hybrid architectures that combine both paradigms.</p>
      <p>Knowledge-based approaches rely on sentiment lexicons and manually curated dictionaries to
determine the polarity of text and words. These methods typically incorporate resources such as
SentiWordNet [14], and domain-specific vocabularies that capture sentiment-bearing words. While interpretable
and linguistically grounded, these approaches often struggle with context-dependent sentiment and
domain-specific expressions common in tourism reviews.</p>
      <p>Machine learning approaches can be further divided into traditional feature engineering methods and
deep learning architectures. Traditional methods focus on extracting handcrafted features such as BOW
countings, TF-IDF vectors, -grams, and part-of-speech patterns, which are then fed to classifiers like
Support Vector Machines, Multinomial Naive Bayes classifiers, or Random Forest [ 15]. Recent advances
have shown that logistic regression with TF-IDF features remains competitive for text classification
tasks, particularly when computational eficiency is prioritized [16].</p>
      <p>The emergence of transformer-based architectures has revolutionized sentiment analysis, with models
like BERT [17] demonstrating superior performance through contextual understanding. Pre-trained
language models, initially developed for English, have been successfully adapted to Spanish through
multilingual variants like XLM-RoBERTa [18, 19] and language-specific models. The recent development
of eficient transformer architectures, such as the Qwen3 family of models [ 20], has made fine-tuning
accessible for resource-constrained scenarios while maintaining competitive performance.</p>
      <p>Hybrid systems combine the interpretability of knowledge-based methods with the learning capacity
of machine learning approaches. Our previous work [21] demonstrated the efectiveness of combining
scored word embeddings with vector representations for tourism sentiment analysis. Similarly, [22]
used word2vec embeddings to construct sentiment dictionaries for social media analysis.</p>
      <p>Tourism-specific sentiment analysis presents unique challenges due to the multilingual nature of
reviews and cultural context dependencies. Recent studies have shown that tourism reviews often
exhibit diferent sentiment patterns compared to general product reviews, with aspects like location,
service quality, and cultural experiences requiring specialized treatment [23]. The emergence of
largescale tourism datasets has enabled more sophisticated modeling approaches, though class imbalance
remains a persistent challenge.</p>
      <p>Named Entity Recognition (NER) has become increasingly important in tourism applications,
particularly for location identification, identifying geographical entities crucial for destination-specific
analysis, and aspect-based sentiment analysis [24, 25]. The integration of NER with sentiment analysis
enables more nuanced understanding of location-specific opinions and experiences.</p>
      <p>Multi-task learning approaches have gained traction in recent years, with architectures designed
to simultaneously predict multiple aspects of text content [26]. Our approach operates within this
paradigm, employing independent specialized parallel models for each prediction target (polarity, type,
and town).</p>
      <p>The Rest-Mex shared task series has provided valuable benchmarks for Spanish tourism sentiment
analysis, with previous editions highlighting the efectiveness of ensemble methods and domain-specific
adaptations [31, 32, 33, 27, 28]. Our current work builds upon these foundations by combining
stateof-the-art transformer fine-tuning with traditional machine learning robustness for comprehensive
tourism review analysis.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methods</title>
      <p>The task addressed in this study involves a multi-label classification problem for Spanish-language
tourism reviews from Mexican destinations. The training dataset consists of 208051 reviews of tourist
destinations across Mexico, structured with six primary columns that capture comprehensive
information about each tourist experience. The Title column contains the brief headline given by tourists to
summarize their opinion, while the Review column includes the full detailed text of their experience. The
dataset includes three target variables for prediction: Polarity, representing sentiment on a five-point
scale from 1 (very negative) to 5 (very positive); Type, categorizing destinations as Hotel, Restaurant, or
Attractive; and Town, identifying the specific location from a list of 40 oficially designated Mexican
Magical Towns. Additionally, the Region column provides the Mexican state information, which serves
as supplementary context rather than a classification target variable.</p>
      <p>The test dataset contains 89166 instances and follows the same structure as the training set, with the
exception that it does not include the Region column, requiring models to rely solely on textual content
and other available features for prediction.</p>
      <p>Each review exhibits typical characteristics of user-generated content, including varied writing styles,
colloquial expressions, and encoding inconsistencies common in Spanish-language web-scraped data.</p>
      <p>In this section, we present our top-performing architecture, which employs a parallel processing
approach for the three aforementioned tasks. Specifically, the system consists of three independent
models, each dedicated to a single task and operating concurrently without inter-model dependencies.
We now describe each of these three models.</p>
      <p>Each document  in the training dataset has the form</p>
      <p>Where the text  represents the text content formed by concatenating the title and opinion fields,
 indicates the regional information and  ,  , and  are the three classification labels Polarity
( ), Type ( ) and Town ( ).</p>
      <sec id="sec-3-1">
        <title>3.1. Preprocessing</title>
        <p>The preprocessing pipeline for this dataset involved several steps. Initially, the provided training dataset
was split into training and testing sets using stratified sampling to maintain class distribution. Missing
values in titles were filled with blank spaces, while duplicate reviews were removed by keeping only
the last occurrence of each duplicated entry.</p>
        <p>Next, character encoding issues were addressed implementing a Latin-1 to UTF-8 conversion process
that successfully corrected 18,451 instances of malformed characters, restoring proper accents and
special characters essential for Spanish language processing. The title and review fields were then
concatenated into a single text feature to capture the complete semantic content of each review.</p>
        <p>From this point, two distinct preprocessing paths were implemented to accommodate diferent
modeling approaches: one tailored for traditional Machine Learning algorithms, and another optimized
for transformer-based models such as BERT, and other modern language models.</p>
        <p>The former version was obtained by using text normalization, which included converting all content
to lowercase and applying tokenization using a blank Spanish language model from SpaCy. The
tokenization process filtered out punctuation, excessive whitespace, and non-alphabetic tokens while
removing Spanish stopwords along with additional noise characters (e.g. \n, \b).</p>
        <p>The latter version of the dataset was obtained by applying minimal preprocessing. Unlike traditional
approaches, this preprocessing pipeline preserved the original text structure and linguistic features
that are crucial for contextual understanding in modern language models. The text underwent basic
cleaning to remove excessive whitespace, and additional noise characters (e.g. \n, \b), while
maintaining punctuation, capitalization patterns, and stopwords that provide important contextual cues for
transformer attention mechanisms [17].</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Polarity Model</title>
        <p>The Polarity Model  is a Large Language Model fine-tuned for 5-class text classification based on
Qwen/Qwen3-0.6B [20], a 0.6-billion-parameter decoder-only transformer originally pretrained for
text generation. The Qwen3 architecture features 28 hidden layers with 16 attention heads and
1024dimensional hidden states, using SiLU activation and RMS layer normalization ( = 1 × 10− 6). To adapt
it to the classification task, we employed the AutoModelForSequenceClassification wrapper
from Hugging Face Transformers, adding a classification head with 5 output labels corresponding to
star ratings (1-5 stars). Since generative models typically lack a padding token, we explicitly added
one to the tokenizer and updated the model configuration accordingly. Input texts were tokenized to
a maximum length of 128 tokens (selected from the original 40960 max_position_embeddings) with
padding and truncation. Fine-tuning was performed using the Hugging Face Trainer API with the
AdamW optimizer, a learning rate of 2 × 10− 5, a batch size of 4, and training over 3 epochs with a
weight decay of 0.01. Evaluation was conducted at the end of each epoch using accuracy and F1 score.</p>
        <p>As shown in Figure 1, the dataset exhibits significant class imbalance, with ratings distributed as: 5
(136,561 reviews, 65.6%), 4 (45,034, 21.6%), 3 (15,519, 7.5%), 2 (5,496, 2.6%), and 1 (5,441, 2.6%). This 25:1
ratio between majority and minority classes necessitated weighted loss functions during training.</p>
        <p>Since the imbalance reflects genuine user behavior patterns, we intentionally preserved the original
distribution during training. This approach maintains the model’s exposure to the natural data
distribution it will encounter during inference [29], while prioritizing performance on majority classes that
dominate real-world use cases. However, to mitigate potential bias while respecting the data’s natural
skew, we employed weighted evaluation metrics and closely monitored per-class accuracy throughout
training [30].</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Type Model</title>
        <p>
          The Type Model  is a logistic regression classifier with 2-regularization. The classifier was trained
on TF-IDF features extracted from the preprocessed text using a TfidfVectorizer with an -gram
range of (
          <xref ref-type="bibr" rid="ref1 ref3">1, 3</xref>
          ), capturing unigrams, bigrams, and trigrams. The vocabulary was limited to the 10,000
most frequent features.
        </p>
        <p>As shown in Figure 2, the type labels exhibit near-balanced distribution, allowing us to employ
standard training protocols without specialized class balancing techniques.</p>
        <p>The logistic regression model was configured with a regularization strength  = 1.0, a tolerance of
10− 4 for the stopping criterion, and a maximum of 1000 iterations.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Town Model</title>
        <p>The Town Model  combines two specialized sub-models, ,1 and ,2, designed to handle
both explicit location mentions and contextual inference for the town classification respectively.</p>
        <p>The sub-model ,1 predicts town labels by matching location entities (LOCs) against a predefined
dictionary {(, {(1), ..., ()})}, achieving 10% coverage. For the remaining 90% of documents lacking
dictionary LOCs, ,2 performs contextual prediction using Machine Learning methods.</p>
      </sec>
      <sec id="sec-3-5">
        <title>3.4.1. Named Entity Recognition Sub-model ,1</title>
        <p>Sub-model ,1 operates as a location mapping dictionary, where:
• Keys: Town labels (e.g., “Dolores Hidalgo”).</p>
        <p>• Values: Associated location named entities learned from training data (e.g., “Atotonilco”).
The sub-model predicts towns by matching extracted named entities in test reviews against this
dictionary. For each review text  , the prediction pred is given by
pred =
{︃
if ∃ ∈  where  ∈ {1 ()</p>
        <p>
          (), . . . ,  }
no prediction, otherwise
(
          <xref ref-type="bibr" rid="ref1">1</xref>
          )
        </p>
        <p>While this mechanism is precise for entries containing known location references, its coverage is
limited because ∼ 90% of test reviews do not contain NERs matching the dictionary.</p>
        <p>We now formally describe the mapping dictionary construction process. The model ,1 is based on
MMG/xlm-roberta-large-ner-spanish, a pretrained XLM-RoBERTa model fine-tuned for Spanish
Named Entity Recognition. For each training document  , we perform inference using this model to
extract a list of named entities (, ), where:
• : The detected entity text.</p>
        <p>• : The entity type (LOC for locations, PER for persons, ORG for organizations, etc.).</p>
        <p>Given our focus on identifying the town class, we retain only geographic entities, i.e. those labeled as
LOC. This step ensures that our model prioritizes location based entities while filtering out irrelevant
entities that could introduce noise.</p>
        <p>We construct a mapping dictionary that associates each town label  with its corresponding location
entities {1 ()</p>
        <p>(), . . . ,  } extracted from training documents labeled as . The dictionary structure
follows the format shown in Table 1, where each entry pairs a town with its distinctive geographical
references.</p>
        <p>As evidenced in Figure 3, several location entities appear in multiple town classes (e.g., “México”).
To ensure unambiguous town prediction, we eliminated these ambiguous entities, creating a reduced
mapping dictionary where each remaining location entity uniquely identifies exactly one town. In
Figure 4 demonstrates this filtering process applied to the three representative cases from Figure 3</p>
        <p>We finally perform town prediction using this reduced mapping dictionary according to Equation 1,
as previously described.</p>
      </sec>
      <sec id="sec-3-6">
        <title>3.4.2. Contextual Prediction Sub-model ,2</title>
        <p>For documents without LOC-based predictions from ,1 (described in Equation 1), the sub-model
,2 employs a two-stage hierarchical classifier to address the 40-class imbalance in town labels  .
The architecture leverages both textual  and regional  features through:
1. Region Classification:
• Input: BERT embeddings of document text 
• Task: Predict region 
• Output: Predicted Region 
• Input: BERT embeddings of document text 
• Task: Predict town  within predicted region 
• Output: Predicted town 
2. Town Classification: There are 12 regional classification models, each one consists of
This hierarchical region-town approach provides three key advantages over direct 40-class town
classification, as evidenced by the distribution patterns in Figures 5 and 6: Hierarchical Class Separation,
Imbalance Mitigation, and Regional Specialization. Each of these advantages is examined below:
• Hierarchical Class Separation: The two-stage architecture decomposes the original 40-class
problem into more manageable sub-tasks. As visible in Figure 5, regional grouping naturally
group towns with linguistically similar vocabulary, while Figure 6 reveals cases where single-town
dominance (e.g., Tulum in Quintana Roo) efectively reduces the classification task to regional
prediction.
• Imbalance Mitigation: The 62:1 imbalance ratio at town level (Figure 5a) is alleviated by grouping
towns into regions, where:
– 8 regions contain ≤ 3 towns (Figure 6).
– Maximum regional imbalance drops to 9:1.</p>
        <p>– 6 regions become binary near-balanced classifiers.
• Regional Specialization: Each regional classifier adapts to local linguistic patterns, avoiding the
noise from nationally dominant tourist vocabularies that would bias a flat classifier.</p>
        <p>Both the region and town classifiers are SVM classifiers, with Gaussian Kernel, taking BERT
embeddings as inputs.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments</title>
      <p>The experimental evaluation assessed model performance across all three classification tasks (polarity,
type, and town), with Weighted macro F1-score as the primary metric to address class imbalance.
All experiments were conducted on a computational node equipped with an NVIDIA L4 GPU (24GB
VRAM) using CUDA 12.2, with PyTorch 1.13.1 and Hugging Face Transformers 4.26.1 libraries. The
hardware-software configuration enabled eficient parallel execution, particularly for the Qwen3-0.6B
ifne-tuning, which required almost 5 hours of training time.</p>
      <p>We adopted a 4:1 train-validation split, stratified by the respective labels, to maintain distributional
consistency. All models were trained on the training subset, with hyperparameters as detailed in Section
3. These hyperparameters were chosen using Grid Search. To ensure reproducibility, we fixed the same
random seeds across all experiments.</p>
      <p>The fine-tuned Qwen3-0.6B model  achieved 73.3% weighted F1-score on the validation dataset,
demonstrating robust performance despite severe class imbalance. The type model  achieved 95.16%
F1-score. Finally, the model  achieved 78% weighted F1-score on the validation dataset across all
town labels.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Results and Discussion</title>
      <p>Our model obtained the results detailed in Table 2 and 7. It ranked almost in the 70th percentile. The
best performing model was the town model .</p>
      <p>As observed in Table 3, our hierarchical model  demonstrated particularly solid performance
in the town classification task, achieving a Macro F1 of 0.6006 that positions it in the 75th percentile.
The detailed analysis by town reveals that the model reached percentiles above 80% in multiple towns,
particularly excelling in towns such as Tepoztlán, Tequila, Cuetzalan, and Tapalpa (88th percentile).</p>
      <p>This consistency in performance across diferent Mexican towns suggests that the model successfully
captured the distinctive linguistic characteristics associated with each town.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusions</title>
      <p>
        In this paper, we proposed a parallel NLP pipeline for sentiment and multi-label classification of Spanish
tourism reviews from Mexican Magical Towns, as part of the Rest-Mex 2025 shared task. Our architecture
combined three specialized models: (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ) a fine-tuned Qwen3-0.6B transformer for 5-class sentiment
analysis, (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) a TF-IDF-based logistic regression classifier for destination type prediction, and (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ) a
Bert-based NER-enhanced hierarchical model for town identification. This modular approach achieved
competitive results, ranking in the 69.57th percentile overall, with particularly strong performance in
town classification (75.36th percentile).
      </p>
      <p>Our key technical contribution is a NER-augmented hierarchical classifier that achieved 60% macro
F1score by combining precise location entity recognition with contextual BERT embeddings, demonstrating
efectiveness for geographically imbalanced datasets (62:1 class ratio).</p>
      <p>The town model’s consistent performance across locations (reaching 88th percentile in Tepoztlán,
Tequila, and Cuetzalan) suggests successful capture of region-specific linguistic patterns. However,
the polarity model’s lower percentile ranking (47.53%) reveals persistent challenges in fine-grained
sentiment analysis for tourism reviews, likely due to subjective labeling and cultural nuances.</p>
      <p>Future work should investigate strategies to better leverage regional information for grouping towns
in ways that enhance classifier performance. Additionally, expanding the scope of entities considered
in the analysis, as well as incorporating town-specific key terms and linguistic markers, could further
improve classification accuracy. Our results confirm that hybrid architectures provide a robust and
efective framework for tourism-oriented NLP applications, especially when computational eficiency
and model transparency are critical requirements.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>The authors acknowledge the High-Performance Computing Area (Área de Cómputo de Alto
Rendimiento, ACARUS) of the University of Sonora for providing the supercomputing infrastructure
essential to this research. https://acarus.unison.mx</p>
    </sec>
    <sec id="sec-8">
      <title>Declaration on Generative AI</title>
      <p>The author(s) have not employed any Generative AI tools.
Roberta: A robustly optimized bert pretraining approach, 2019. URL: https://arxiv.org/abs/1907.
11692. arXiv:1907.11692.
[20] A. Yang, A. Li, B. Yang, B. Zhang, B. Hui, B. Zheng, B. Yu, C. Gao, C. Huang, C. Lv, C. Zheng, D. Liu,
F. Zhou, F. Huang, F. Hu, H. Ge, H. Wei, H. Lin, J. Tang, J. Yang, J. Tu, J. Zhang, J. Yang, J. Yang,
J. Zhou, J. Zhou, J. Lin, K. Dang, K. Bao, K. Yang, L. Yu, L. Deng, M. Li, M. Xue, M. Li, P. Zhang,
P. Wang, Q. Zhu, R. Men, R. Gao, S. Liu, S. Luo, T. Li, T. Tang, W. Yin, X. Ren, X. Wang, X. Zhang,
X. Ren, Y. Fan, Y. Su, Y. Zhang, Y. Zhang, Y. Wan, Y. Liu, Z. Wang, Z. Cui, Z. Zhang, Z. Zhou, Z. Qiu,
Qwen3 technical report, 2025. URL: https://arxiv.org/abs/2505.09388. arXiv:2505.09388.
[21] M. Toledo-Acosta, T. Barreiro, A. Reig-Alamillo, M. Müller, F. Aroca Bisquert, M. L. Barrigon,
E. Baca-Garcia, J. Hermosillo-Valadez, Cognitive emotional embedded representations of text to
predict suicidal ideation and psychiatric symptoms, Mathematics 8 (2020). URL: https://www.mdpi.
com/2227-7390/8/11/2088. doi:10.3390/math8112088.
[22] B. Shi, J. Zhao, K. Xu, A word2vec model for sentiment analysis of weibo, in: 2019 16th International
Conference on Service Systems and Service Management (ICSSSM), 2019, pp. 1–6. doi:10.1109/
ICSSSM.2019.8887652.
[23] S. Anis, S. Saad, M. Aref, Sentiment analysis of hotel reviews using machine learning techniques,
in: A. E. Hassanien, A. Slowik, V. Snášel, H. El-Deeb, F. M. Tolba (Eds.), Proceedings of the
International Conference on Advanced Intelligent Systems and Informatics 2020, Springer International
Publishing, Cham, 2021, pp. 227–234.
[24] B. Cowan, S. Zethelius, B. Luk, T. Baras, P. Ukarde, D. Zhang, Named entity recognition in
travel-related search queries, Proceedings of the AAAI Conference on Artificial Intelligence 29
(2015) 3935–3941. URL: https://ojs.aaai.org/index.php/AAAI/article/view/19050. doi:10.1609/
aaai.v29i2.19050.
[25] D. H. Fudholi, A. Zahra, S. Rani, S. N. Huda, I. V. Paputungan, Z. Zukhri, Bert-based tourism named
entity recognition: making use of social media for travel recommendations, PeerJ Computer
Science 9 (2023) e1731. doi:10.7717/peerj-cs.1731.
[26] S. Ruder, An overview of multi-task learning in deep neural networks, 2017. URL: https://arxiv.</p>
      <p>org/abs/1706.05098. arXiv:1706.05098.
[27] J. Á. González-Barba, L. Chiruzzo, S. M. Jiménez-Zafra, Overview of IberLEF 2025: Natural
Language Processing Challenges for Spanish and other Iberian Languages, in: Proceedings of the
Iberian Languages Evaluation Forum (IberLEF 2025), co-located with the 41st Conference of the
Spanish Society for Natural Language Processing (SEPLN 2025), CEUR-WS. org, 2025.
[28] M. Á. Álvarez-Carmona, Á. Díaz-Pacheco, R. Aranda, A. Y. Rodríguez-González, L. Bustio-Martínez,
V. Herrera-Semenets, Overview of rest-mex at iberlef 2025: Researching sentiment evaluation in
text for mexican magical towns, volume 75, 2025.
[29] M. Buda, A. Maki, M. A. Mazurowski, A systematic study of the class imbalance problem in
convolutional neural networks, Neural Networks 106 (2018) 249–259. URL: https://www.sciencedirect.com/
science/article/pii/S0893608018302107. doi:https://doi.org/10.1016/j.neunet.2018.07.
011.
[30] S. Das, S. S. Mullick, I. Zelinka, On supervised class-imbalanced learning: An updated perspective
and some key challenges, IEEE Transactions on Artificial Intelligence 3 (2022) 973–993. doi: 10.
1109/TAI.2022.3160658.
[31] M. Á. Álvarez-Carmona, R. Aranda, S. Arce-Cárdenas, D. Fajardo-Delgado, R. Guerrero-Rodríguez,
A. P. López-Monroy, J. Martínez-Miranda, H. Pérez-Espinosa, A. Rodríguez-González, Overview
of rest-mex at iberlef 2021: Recommendation system for text mexican tourism, Procesamiento del
Lenguaje Natural 67 (2021). doi:https://doi.org/10.26342/2021-67-14.
[32] M. Á. Álvarez-Carmona, Á. Díaz-Pacheco, R. Aranda, A. Y. Rodríguez-González, D. Fajardo-Delgado,
R. Guerrero-Rodríguez, L. Bustio-Martínez, Overview of rest-mex at iberlef 2022:
Recommendation system, sentiment analysis and covid semaphore prediction for mexican tourist texts,
Procesamiento del Lenguaje Natural 69 (2022).
[33] M. Á. Álvarez-Carmona, Á. Díaz-Pacheco, R. Aranda, A. Y. Rodríguez-González, L. Bustio-Martínez,
V. Muñis-Sánchez, A. P. Pastor-López, F. Sánchez-Vega, Overview of rest-mex at iberlef 2023:
Research on sentiment analysis task for mexican tourist texts, Procesamiento del Lenguaje Natural
71 (2023).
[34] M. Á. Álvarez-Carmona, R. Aranda, A. Y. Rodríguez-González, L. Pellegrin, C. Hugo, Classifying
the mexican epidemiological semaphore colour from the covid-19 text spanish news, Journal of
Information Science (2022). doi:https://doi.org/10.1177/01655515221100952.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Afren</surname>
          </string-name>
          ,
          <article-title>The role of digital marketing promoting tourism business. a study of use of the social media in prompting travel</article-title>
          .,
          <source>World Journal of Advanced Research and Reviews</source>
          <volume>21</volume>
          (
          <year>2024</year>
          )
          <fpage>272</fpage>
          -
          <lpage>287</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Diaz-Pacheco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Álvarez-Carmona</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-GonzÁlez</surname>
          </string-name>
          , H. Carlos,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <article-title>Measuring the diference between pictures from controlled and uncontrolled sources to promote a destination. a deep learning approach</article-title>
          ,
          <source>International Journal of Interactive Multimedia and Artificial Intelligence</source>
          In Press (
          <year>2023</year>
          )
          <fpage>1</fpage>
          -
          <lpage>14</lpage>
          . URL: http://dx.doi.org/10.9781/ijimai.
          <year>2023</year>
          .
          <volume>10</volume>
          .003. doi:
          <volume>10</volume>
          .9781/ijimai.
          <year>2023</year>
          .
          <volume>10</volume>
          .003.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>E.</given-names>
            <surname>Olmos-Martínez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Álvarez-Carmona</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Aranda</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Díaz-Pacheco</surname>
          </string-name>
          ,
          <article-title>What does the media tell us about a destination? the cancun case, seen from the usa, canada, and mexico</article-title>
          ,
          <source>International Journal of Tourism Cities</source>
          <volume>10</volume>
          (
          <year>2023</year>
          )
          <fpage>639</fpage>
          -
          <lpage>661</lpage>
          . URL: http://dx.doi.org/10.1108/IJTC-09-2022-0223. doi:
          <volume>10</volume>
          .1108/ijtc-09-2022-0223.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Álvarez-Carmona</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          , et al.,
          <article-title>Big data analytics of online news to explore destination image using a comprehensive deep-learning approach: a case from mexico</article-title>
          ,
          <source>Information Technology &amp; Tourism</source>
          <volume>26</volume>
          (
          <year>2024</year>
          )
          <fpage>147</fpage>
          -
          <lpage>182</lpage>
          . URL: https://doi.org/10.1007/ s40558-023-00278-5. doi:
          <volume>10</volume>
          .1007/s40558-023-00278-5.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>