<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Hierarchical Attention Networks for Multilabel Sentiment Analysis in Spanish Reviews of Mexican Magic Towns</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ezau Faridh Torres Torres</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Edison David Serrano Cárdenas</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Mathematics Research Center (CIMAT)</institution>
          ,
          <addr-line>Jalisco S/N, Valenciana, 36023 Guanajuato, GTO</addr-line>
          <country country="MX">México</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Secretaría de Ciencia, Humanidades, Tecnología e Innovación (SECIHTI)</institution>
          ,
          <addr-line>Av. de los Insurgentes Sur 1582, Benito Juárez, 03940, CDMX</addr-line>
          ,
          <country country="MX">México</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>In this study, we address the task of joint classification of sentiment polarity, destination type, and Magic Town from Spanish-language tourist reviews, as proposed by the Rest-Mex 2025 challenge, part of the IberLEF 2025 evaluation forum [1, 2]. We hereby propose a lightweight, hierarchical attention-based model that leverages the inherent structure of review data. Specifically, the model posits that words form reviews, and reviews are grouped by town. The proposed approach is predicated on the Hierarchical Attention Network (HAN), with extensions to a multi-output setting that are concomitant with the preservation of eficiency and interpretability. Experimental results show that our model significantly outperforms the oficial baseline and the average of all participating systems, achieving a joint track score of 0.601, with macro-F1 scores of 0.541 for polarity, 0.943 for destination type, and 0.528 for town classification. It is noteworthy that the model demonstrates robust generalization capabilities, particularly in underrepresented classes and low-resource towns. In contrast to large-scale transformer-based models, our architecture can be trained eficiently without the use of GPUs or extensive pretraining, making it suitable for deployment in environments with limited resources.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;REST-MEX 2025</kwd>
        <kwd>Sentiment Analysis</kwd>
        <kwd>Hierarchical Attention Network (HAN)</kwd>
        <kwd>Review Classification</kwd>
        <kwd>Multitask Learning</kwd>
        <kwd>Spanish NLP</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Conventional review classification methodologies employ sparse lexical features, such as -grams,
as review representations, subsequently utilizing either a linear model or kernel methods on these
representations [
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ]. Recent approaches have incorporated deep learning techniques, particularly
transformer-based architectures, such as BERT [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], RoBERTa [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], and its multilingual extension XLM-R
[7]. For Spanish-language applications, specialized models like RoBERTuito [8] have demonstrated
competitive performance in review classification and sentiment analysis tasks.
      </p>
      <p>
        This study was conducted in the context of the Rest-Mex 2025 challenge [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], part of the IberLEF 2025
evaluation forum [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Unlike past editions [9, 10, 11], this year focuses on the joint classification of
tourist review sentiment, destination type, and Magic Town identification from Spanish-language texts.
      </p>
      <p>To address this task, we adapted the Hierarchical Attention Network (HAN) proposed by Yang et al.
[12] to our classification task. In contrast to the original document-level configuration, we categorize
reviews by Magic Town, implementing an approach that encompasses both the word and review levels.</p>
      <p>Recent eforts have explored the use of compact transformer models such as DistilBERT [ 13] to
mitigate the computational demands of full-sized language models like BERT and RoBERTa. While
these approaches reduce parameter counts, they still rely on large-scale pretraining and often require
GPU-intensive resources during fine-tuning and inference. In contrast, our proposed model leverages a
lightweight, interpretable hierarchical architecture based on bidirectional GRUs and attention
mechanisms. This design enables eficient training and robust performance, even in resource-constrained
environments, without the need for massive pretrained language models.</p>
    </sec>
    <sec id="sec-2">
      <title>1. Problem Statement and Task Objectives</title>
      <p>This section outlines the challenges addressed in the “Rest Mex 2025” sentiment analysis competition,
which centers on understanding tourist perceptions through the analysis of user-generated reviews
[14, 15, 16].
1.1. Task
The core objective of the competition is to analyze reviews and opinions related to tourist destinations
in Mexico and extract relevant sentiment-based and categorical information. Specifically, the task is
structured around three main classification goals:
• Sentiment Polarity Classification: Given a textual review, the first objective is to predict its
sentiment on a five-point scale, where a score of 1 denotes strong dissatisfaction and a score of 5
indicates high satisfaction. This allows for a nuanced understanding of tourist sentiment beyond
binary positive/negative classifications.
• Tourist Site Type Classification: The second goal is to categorize the type of site referenced in
the review. The possible categories are hotel, restaurant, or attraction. This classification aids in
diferentiating sentiment patterns across various tourism sectors.
• Magic Town Identification: The final objective is to determine which of the 40 designated
Mexican Magic Town the review refers to. By leveraging metadata and contextual information
within the text, this classification can reveal regional trends and cultural variations in how tourists
express sentiment.</p>
      <p>Each review serves as a valuable source of information about traveler experiences. The overarching
aim is to extract structured and meaningful insights from these texts to better understand tourist
satisfaction, preferences, and regional diferences across Mexico’s culturally rich destinations.</p>
    </sec>
    <sec id="sec-3">
      <title>2. Methodology</title>
      <sec id="sec-3-1">
        <title>2.1. Exploratory Data Analysis</title>
        <p>The dataset provided for the competition consists of 208,051 labeled reviews from tourists sharing their
experiences across various destinations. Each review contains five key fields: a title, the main review
text, the type of place being reviewed, the sentiment polarity on a scale from 1 to 5, and the name of one
of 40 towns in Mexico. Notably, there was only one instance with a missing title, which was handled
during preprocessing.</p>
        <p>A critical part of exploratory analysis involved assessing the class distribution to determine whether
the dataset was balanced. Balanced datasets are essential for ensuring that the model can learn equally
well across all classes. In contrast, imbalanced datasets can lead to biased predictions, especially for
minority classes that are underrepresented during training.</p>
        <p>As shown in Figure 2, the sentiment polarity distribution is heavily skewed toward the most positive
sentiment. Reviews with a polarity score of 5 account for approximately 65.64% of the dataset, indicating
a strong bias toward favorable opinions. This imbalance could potentially afect the model’s ability to
accurately predict lower sentiment scores, as it is exposed to far fewer negative or neutral examples.</p>
        <p>Similarly, the distribution of reviews across towns is notably uneven. As illustrated in Figure 1,
approximately 41.33% of the reviews come from towns located in Quintana Roo, indicating a regional
concentration of tourist activity or review behavior. This imbalance may impact the model’s performance
on towns with fewer samples, as the model could become biased toward the characteristics of more
frequently represented locations.</p>
        <p>In contrast, the distribution of the type of place (i.e., attraction, hotel, or restaurant) is more balanced,
though not perfectly uniform. Specifically, restaurants are the most frequently reviewed category,
representing 41.68% of the dataset, followed by hotels with 24.71%, and attractions, which are
significantly underrepresented at only 0.33%. While this indicates a moderate imbalance, the disparity is less
pronounced than in the sentiment and town distributions.</p>
      </sec>
      <sec id="sec-3-2">
        <title>2.2. Dataset Construction</title>
        <p>The dataset preparation phase was crucial to ensure compatibility of each instance with the requirements
of the sentiment analysis task. Categorical values in key columns were transformed into numerical
representations. Specifically, the Type column was encoded as follows: 0 for Attraction, 1 for Hotel,
and 2 for Restaurant. Similarly, the Town column was mapped to integers ranging from 0 to 39, based
on the alphabetical order of the 40 magic towns included in the dataset.</p>
        <p>To better capture both the context and summary of each review, the title and review columns were
concatenated into a single text field. In cases where the title was missing, it was replaced with an empty
string. This merged column helps retain both the concise summary typically found in the title and the
detailed feedback from the full review.</p>
        <p>Preprocessing steps included converting all text to lowercase and performing cleaning operations to
correct patterns that attempted to represent tildes (e.g., replacing corrupted or miswritten characters
common in Spanish). Given the importance of emoticons in expressing sentiment, various facial
expressions were replaced with corresponding descriptive words. For example, happy faces such as
“:)” and “:))” were replaced with “feliz” (happy), while sad expressions like “:(” were substituted with
“triste” (sad). A broader set of emoticons was considered in this substitution process to better preserve
the emotional content of the reviews. Additionally, common stopwords and non-informative symbols
were removed to reduce noise and improve model focus on relevant content.</p>
        <p>The dataset was shufled to ensure a randomized distribution of instances. For the experimental
setup, it was divided into training and test subsets: 80% of the data (166,440 instances) was allocated for
training, while the remaining 20% (41,611 instances) was reserved for evaluation.</p>
        <p>To address class imbalance during training—particularly for underrepresented sentiment classes
(those with polarity ≤ 3) and towns with fewer than 2,800 reviews—a contextual data augmentation
strategy was applied. This involved in-place token repetition, where a randomly selected segment of
ifve consecutive tokens from the original review was appended to the end of the text. This technique
aimed to enhance the context for minority classes while preserving the natural semantics of the reviews.</p>
        <p>For textual representation, we employed pretrained Word2Vec embedding [17], which provided
semantic-rich vector representations of the Spanish text to feed into our model.</p>
      </sec>
      <sec id="sec-3-3">
        <title>2.3. Hierarchical attention networks</title>
        <p>In this work, we build on the idea that improved review representations can be obtained by incorporating
knowledge of review structure into a hierarchical model architecture. The fundamental intuition
informing our model is that not all components of a review or the collection of reviews linked to a
given Magic Town are equally pertinent to the classification task, and that determining the relevant
sections entails modeling interactions among words, rather than considering their isolated presence.
This idea builds upon prior work on attention-based models [12].</p>
        <p>The objective of the Hierarchical Attention Network (HAN) is to glean two fundamental insights
concerning the structural intricacies of review data. First, given the hierarchical nature of our dataset—where
individual words form reviews, and reviews are grouped by Magic Town—a final representation for each
town is constructed by first encoding each review and then aggregating them into a comprehensive
representation at the town level. Secondly, it has been observed that the informative value of words
within reviews, as well as reviews within a town, varies depending on the classification task. In
particular, the importance of specific words or reviews may change depending on their context [ 12]. To
account for this, the model employs two levels of attention mechanisms [18, 19]: one at the word level
and another at the review level. These mechanisms allow the model to dynamically adjust the weight
assigned to each component during the construction of the town-level representation.</p>
        <p>
          To illustrate this phenomenon, consider the example in Figure 4, which is a brief Yelp review wherein
the objective is to predict the rating on a scale from 1 − 5 [20]. Intuitively, the first and third sentences
appear to contain more pertinent information for predicting the rating. Within these sentences, the
words "delicious" and "a-m-a-z-i-n-g", which are frequently used to express a positive sentiment,
contributes to the implication of the positive attitude contained in this review. Attention confers two
advantages: it frequently leads to enhanced performance and it ofers insight into which words and
sentences contribute to the classification decision. This insight can be valuable in applications and
analysis [
          <xref ref-type="bibr" rid="ref4">4, 21</xref>
          ].
        </p>
        <p>The overall architecture of the Hierarchical Attention Network (HAN) is illustrated in Figure 3. The
model is composed of multiple components, including a word sequence encoder, a word-level attention
layer, a review encoder, and a review-level attention layer.</p>
        <p>Our approach aligns with the hierarchical attention formulation proposed by Yang et al. [12], which
involves the acquisition of both word-level and review-level representations through a learned attention
mechanism:</p>
        <p>exp (⊤)
  = ∑︀ exp (⊤)
(1)
In our implementation, both word and review sequences are encoded using bidirectional Gated Recurrent
Units (GRUs), which allow the model to capture contextual dependencies from both past and future
tokens or reviews. At each level, the GRU outputs a sequence of hidden states ℎ, which are then
processed by an attention mechanism. Here, ℎ denotes the hidden state at time step , and  is
a learned context vector used to compute the importance weights   via a softmax operation. The
resulting vector  is a weighted sum that encodes the most informative parts of the sequence. This
formulation is applied at two distinct levels: first, at the word level, to obtain review embeddings; and
second, at the review level, to obtain town embeddings.</p>
        <p>Unlike the original HAN designed for document classification, our model aggregates multiple reviews
associated with a Magic Town, and predicts multiple outputs at diferent levels (per-review and
pertown). We preserve the hierarchical attention structure but adapt it to a multitask scenario.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>3. Results and discussion</title>
      <sec id="sec-4-1">
        <title>3.1. Evaluation</title>
        <p>The evaluation section is responsible for assessing the performance and efectiveness of the proposed
model across three classification tasks: polarity ( 1 − 5), type (Attractive, Hotel, Restaurant), and Magic
Town (40 classes). The results obtained by our best submission, FrogCode_1, are reported herein.
FrogCode_1 consistently outperformed the provided baseline across all metrics.</p>
        <p>The evaluation process utilizes Macro-F1 scores for each task, aligning with the oficial metrics
employed in the Rest-Mex 2025 competition. In addition to aggregate performance (Track Score), we
analyze per-class F1, precision, and recall to gain further insights into the model’s behavior.</p>
        <p>In order to provide a comprehensive evaluation of the results, comparative analyses with the oficial
baseline and statistical summaries across all submissions are also included. These additional analyses
are used to contextualize the performance in question.</p>
      </sec>
      <sec id="sec-4-2">
        <title>3.2. Competition results</title>
        <p>As illustrated in Table 1, the model’s final performance metrics are presented in comparison to the
oficial baseline and the overall average across all participating systems in the Rest-Mex 2025 Sentiment
Analysis track. The model demonstrated a Track Score of 0.601, which is significantly higher than the
baseline (0.090) and the average participant score (0.484). It is noteworthy that the model achieved
Macro F1 scores of 0.541, 0.943, and 0.528 for polarity, type, and town classification, respectively. These
scores surpassed the mean scores and the baseline by a substantial margin. These results demonstrate
the efectiveness of incorporating a hierarchical attention mechanism for modeling review structure,
particularly in the town-level classification task, which exhibited the largest gap over the baseline
(0.528 vs. 0.009).</p>
        <p>Model
Baseline
Overall average
Our model</p>
        <p>Track Score</p>
        <p>Macro F1 (Polarity)</p>
        <p>Macro F1 (Type)</p>
        <p>Macro F1 (Town)
0.090
0.484
0.601
0.158
0.449
0.541
0.197
0.796
0.943
0.009
0.403
0.528</p>
        <p>In addition to the aggregated metrics presented in Table 1, it was observed that there was consistent
performance across all subtasks, including underrepresented and imbalanced classes. Notwithstanding
the considerable polarity skew (with over 136, 000 reviews rated as 5), the model exhibited competitive
Macro F1 scores of 0.584 for polarity 1, 0.333 for polarity 2, and 0.483 for polarity 3. This finding
indicates that the architecture did not merely adapt to the majority class, but rather, it efectively
captured informative patterns across the sentiment spectrum.</p>
        <p>In the context of the type classification task, the model demonstrated notable precision and
generalization capabilities, achieving F1 scores of 0.95, 0.93, and 0.95 for the categories "Attractive," "Hotel,"
and "Restaurant," respectively, even when confronted with varying data volumes.</p>
        <p>In addition, the model’s notable robustness was demonstrated in the town classification task, with F1
scores exceeding 0.73 for towns with substantial data, such as Isla Mujeres, Teotihuacan, and Tulum.
Notably, the model demonstrated notable performance in low-resource towns, as evidenced by its F1
values of 0.227 in Tepotzotlan, 0.345 in Cuetzalan, and 0.401 in Metepec and as we have seen in Seccion
2.1. This finding underscores the model’s capacity for generalization across long-tail distributions.</p>
      </sec>
      <sec id="sec-4-3">
        <title>3.3. Error Analysis and Challenging Cases</title>
        <p>An analysis of the confusion matrix on the validation set reveals notable challenges in classifying
certain towns. Tepotzotlán, Cuetzalan, and Metepec exhibit high misclassification rates. Tepotzotlán
shows a low number of correct predictions and widespread confusion with other towns, indicating
weak feature representation. Cuetzalan’s errors are broadly distributed across multiple classes, while
Metepec, despite better accuracy, is frequently confused with semantically similar locations.</p>
        <p>Moreover, misclassifications tend to follow regional patterns. Towns from the same state—such as
Tepotzotlán, Metepec, and Valle de Bravo (State of Mexico)—are often confused. A similar trend is
observed among geographically close towns like Valladolid (Yucatán), Tulum, and Isla Mujeres (Quintana
Roo). This suggests that tourist reviews from nearby or culturally similar towns share linguistic features
that make them harder to distinguish.</p>
        <p>These findings highlight the need for improved handling of class imbalance and the incorporation of
regional and semantic context to enhance model performance in town-level classification.</p>
      </sec>
      <sec id="sec-4-4">
        <title>3.4. Comparative Discussion with Transformer Models</title>
        <p>Although transformer-based architectures such as BERT and RoBERTa have demonstrated
state-ofthe-art performance in various NLP tasks, they often require substantial computational resources and
extensive pretraining. In contrast, our hierarchical attention-based model achieves competitive results
with significantly fewer parameters and without relying on transfer learning.</p>
        <p>Recent studies have proposed lightweight transformer variants, such as DistilBERT [13] and
TinyBERT [22], to alleviate some of these limitations. However, even these models entail GPU-intensive
ifne-tuning and storage of pretrained weights. Our model, by leveraging a hierarchical BiGRU-attention
mechanism, delivers a joint Track Score of 0.601 and consistently strong macro-F1 scores, without
incurring such computational costs.</p>
        <p>This comparison highlights that HAN-based architectures remain viable for structured classification
tasks, particularly in low-resource settings where transformer-based approaches may be impractical.
Future research could explore hybridizing HANs with distilled transformers to balance interpretability,
eficiency, and raw performance.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>4. Conclusions</title>
      <p>In this work, we proposed a hierarchical attention-based model tailored for the Rest-Mex 2025 challenge,
addressing the joint classification of sentiment polarity, type of destination, and Magic Town based on
Spanish-language tourist reviews. The model’s eficacy in capturing multilevel contextual dependencies
is attributable to its utilization of the hierarchical structure of the data, wherein words are classified
as reviews and reviews are aggregated by town. The dual attention mechanism enabled the model to
dynamically focus on the most informative words and reviews, resulting in robust performance across
all tasks.</p>
      <p>The eficacy of our approach was found to exceed not only the oficial baseline but also the mean
performance of the systems that were engaged in the study. This performance was evidenced by the
attainment of a Track Score of 0.601 and the consistent generation of elevated Macro F1 scores across the
entirety of the subtasks that were utilized in the evaluation process. Notably, the model demonstrated
strong generalization on underrepresented classes and low-resource towns, showcasing its adaptability
to real-world data imbalances.</p>
      <p>Furthermore, the model’s lightweight architecture ofers practical advantages in real-world scenarios.
It can be trained and deployed eficiently on standard hardware without reliance on GPU acceleration
or large-scale pretraining. This makes it particularly well-suited for use in settings with limited
computational resources, such as public sector institutions, small tourism operators, or NGOs working
with regional data. Its interpretable design also facilitates model transparency, an important factor for
decision-making in applied contexts.</p>
      <p>These results underscore the eficacy of hierarchical architectures in multi-output text classification
settings, particularly when confronted with nested, skewed, and multilingual datasets. Subsequent
studies may augment this framework by incorporating multilingual pretraining or uncertainty estimation
to enhance its robustness and interpretability.</p>
      <p>The experimentation can be found in a public GitHub repository [23].</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>We declare that the present manuscript has been written entirely by the authors and that no generative
artificial intelligence tools were used in its preparation, drafting, or editing.
Roberta: A robustly optimized bert pretraining approach, 2019. URL: https://arxiv.org/abs/1907.
11692. arXiv:1907.11692.
[7] A. Conneau, K. Khandelwal, N. Goyal, V. Chaudhary, G. Wenzek, F. Guzmán, E. Grave, M. Ott,
L. Zettlemoyer, V. Stoyanov, Unsupervised cross-lingual representation learning at scale, 2020.</p>
      <p>URL: https://arxiv.org/abs/1911.02116. arXiv:1911.02116.
[8] J. M. Pérez, D. A. Furman, L. A. Alemany, F. Luque, Robertuito: a pre-trained language model for
social media text in spanish, 2022. URL: https://arxiv.org/abs/2111.09453. arXiv:2111.09453.
[9] M. Á. Álvarez-Carmona, R. Aranda, S. Arce-Cárdenas, D. Fajardo-Delgado, R. Guerrero-Rodríguez,
A. P. López-Monroy, J. Martínez-Miranda, H. Pérez-Espinosa, A. Rodríguez-González, Overview
of rest-mex at iberlef 2021: Recommendation system for text mexican tourism, Procesamiento del
Lenguaje Natural 67 (2021). doi:https://doi.org/10.26342/2021-67-14.
[10] M. Á. Álvarez-Carmona, Á. Díaz-Pacheco, R. Aranda, A. Y. Rodríguez-González, D. Fajardo-Delgado,
R. Guerrero-Rodríguez, L. Bustio-Martínez, Overview of rest-mex at iberlef 2022:
Recommendation system, sentiment analysis and covid semaphore prediction for mexican tourist texts,
Procesamiento del Lenguaje Natural 69 (2022).
[11] M. Á. Álvarez-Carmona, Á. Díaz-Pacheco, R. Aranda, A. Y. Rodríguez-González, L. Bustio-Martínez,
V. Muñis-Sánchez, A. P. Pastor-López, F. Sánchez-Vega, Overview of rest-mex at iberlef 2023:
Research on sentiment analysis task for mexican tourist texts, Procesamiento del Lenguaje Natural
71 (2023).
[12] Z. Yang, D. Yang, C. Dyer, X. He, A. Smola, E. Hovy, Hierarchical attention networks for document
classification, in: NAACL 2016, 2016, pp. 1480–1489. URL: https://www.microsoft.com/en-us/
research/publication/hierarchical-attention-networks-document-classification/.
[13] V. Sanh, L. Debut, J. Chaumond, T. Wolf, Distilbert, a distilled version of bert: smaller, faster,
cheaper and lighter, 2020. URL: https://arxiv.org/abs/1910.01108. arXiv:1910.01108.
[14] R. Guerrero-Rodriguez, M. A. Álvarez Carmona, R. Aranda, A. P. López-Monroy,
Studying online travel reviews related to tourist attractions using nlp methods: the
case of guanajuato, mexico, Current Issues in Tourism 26 (2023) 289–304. URL:
https://doi.org/10.1080/13683500.2021.2007227. doi:10.1080/13683500.2021.2007227.
arXiv:https://doi.org/10.1080/13683500.2021.2007227.
[15] M. A. Álvarez-Carmona, R. Aranda, A. Y. Rodríguez-Gonzalez, D. Fajardo-Delgado, M. G. Sánchez,
H. Pérez-Espinosa, J. Martínez-Miranda, R. Guerrero-Rodríguez, L. Bustio-Martínez, Ángel
DíazPacheco, Natural language processing applied to tourism research: A systematic review and future
research directions, Journal of King Saud University - Computer and Information Sciences 34
(2022) 10125–10144. URL: https://www.sciencedirect.com/science/article/pii/S1319157822003615.
doi:https://doi.org/10.1016/j.jksuci.2022.10.010.
[16] E. Olmos-Martínez, M. Á. Álvarez-Carmona, R. Aranda, A. Díaz-Pacheco, What does the media
tell us about a destination? the cancun case, seen from the usa, canada, and mexico, International
Journal of Tourism Cities 10 (2023) 639–661. URL: http://dx.doi.org/10.1108/IJTC-09-2022-0223.
doi:10.1108/ijtc-09-2022-0223.
[17] T. Mikolov, K. Chen, G. Corrado, J. Dean, Eficient estimation of word representations in vector
space, 2013. URL: https://arxiv.org/abs/1301.3781. arXiv:1301.3781.
[18] D. Bahdanau, K. Cho, Y. Bengio, Neural machine translation by jointly learning to align and
translate, 2016. URL: https://arxiv.org/abs/1409.0473. arXiv:1409.0473.
[19] K. Xu, J. Ba, R. Kiros, K. Cho, A. Courville, R. Salakhutdinov, R. Zemel, Y. Bengio, Show, attend
and tell: Neural image caption generation with visual attention, 2016. URL: https://arxiv.org/abs/
1502.03044. arXiv:1502.03044.
[20] N/A, Yelp 2013, 2023. URL: https://doi.org/10.5281/zenodo.7555898. doi:10.5281/zenodo.</p>
      <p>7555898.
[21] J. Gao, P. Pantel, M. Gamon, X. He, L. Deng, Modeling interestingness with deep neural networks, in:
A. Moschitti, B. Pang, W. Daelemans (Eds.), Proceedings of the 2014 Conference on Empirical
Methods in Natural Language Processing (EMNLP), Association for Computational Linguistics, Doha,
Qatar, 2014, pp. 2–13. URL: https://aclanthology.org/D14-1002/. doi:10.3115/v1/D14-1002.
[22] X. Jiao, Y. Yin, L. Shang, X. Jiang, X. Chen, L. Li, F. Wang, Q. Liu, Tinybert: Distilling bert for natural
language understanding, 2020. URL: https://arxiv.org/abs/1909.10351. arXiv:1909.10351.
[23] E. Serrano, E. Torres, NLP_Rest_Mex2025, https://github.com/edserranoc/NLP_Rest_Mex2025,
2025. Accessed: 2025-05-21.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <article-title>Álvarez-Carmona, Á</article-title>
          . Díaz-Pacheco,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bustio-Martínez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Herrera-Semenets</surname>
          </string-name>
          ,
          <article-title>Overview of rest-mex at iberlef 2025: Researching sentiment evaluation in text for mexican magical towns</article-title>
          , volume
          <volume>75</volume>
          ,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>González-Barba</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Chiruzzo</surname>
            ,
            <given-names>S. M.</given-names>
          </string-name>
          <string-name>
            <surname>Jiménez-Zafra</surname>
          </string-name>
          ,
          <article-title>Overview of IberLEF 2025: Natural Language Processing Challenges for Spanish and other Iberian Languages, in: Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2025), co-located with the 41st Conference of the Spanish Society for Natural Language Processing (SEPLN 2025), CEUR-WS</article-title>
          . org,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Manning</surname>
          </string-name>
          ,
          <article-title>Baselines and bigrams: Simple, good sentiment and topic classification</article-title>
          , in: H.
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>C.-Y.</given-names>
          </string-name>
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Osborne</surname>
            ,
            <given-names>G. G.</given-names>
          </string-name>
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>J. C.</given-names>
          </string-name>
          Park (Eds.),
          <source>Proceedings of the 50th Annual Meeting of the Association for Computational Linguistics (Volume</source>
          <volume>2</volume>
          :
          <string-name>
            <surname>Short</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>Association for Computational Linguistics</source>
          , Jeju Island, Korea,
          <year>2012</year>
          , pp.
          <fpage>90</fpage>
          -
          <lpage>94</lpage>
          . URL: https://aclanthology.org/ P12-2018/.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>T.</given-names>
            <surname>Joachims</surname>
          </string-name>
          ,
          <article-title>Text categorization with support vector machines: Learning with many relevant features</article-title>
          , in: C.
          <string-name>
            <surname>Nédellec</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          Rouveirol (Eds.),
          <source>Machine Learning: ECML-98</source>
          , Springer Berlin Heidelberg, Berlin, Heidelberg,
          <year>1998</year>
          , pp.
          <fpage>137</fpage>
          -
          <lpage>142</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , BERT:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          , in: J.
          <string-name>
            <surname>Burstein</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Doran</surname>
          </string-name>
          , T. Solorio (Eds.),
          <source>Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>1</volume>
          (Long and Short Papers),
          <source>Association for Computational Linguistics</source>
          , Minneapolis, Minnesota,
          <year>2019</year>
          , pp.
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          . URL: https://aclanthology.org/N19-1423/. doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>N19</fpage>
          -1423.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Joshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          ,
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>