<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Natural Language Processing Approaches for Spanish Tourist Review Analysis: Insights from the REST-MEX 2025 Shared Task</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Alvaro Cabrera-Barrio</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Yuri Melissa Garcia-Niño</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Victor Alfredo Apaza-Mamani</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Universidad Carlos III de Madrid, Av. de la Universidad</institution>
          ,
          <addr-line>30, 28911 Leganes, Madrid</addr-line>
          ,
          <country>España</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>In this paper, we describe our participation in the IberLEF 2025 shared task on analyzing TripAdvisor reviews of Mexican Magical Towns. The goal is to classify each review by sentiment polarity from 1 to 5, type of site (hotel, restaurant, or attraction), and the associated Magical Town from a predefined list. To address these subtasks, we explored several NLP techniques for Spanish language input, including Support Vector Machines (SVM), finetuned transformer models and prompting strategies with large language models. Our models were good but hindered by the dificult task of classifying between 40 diferent Magical Towns, as we will further develop in this document.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Sentiment Analysis</kwd>
        <kwd>Text classification</kwd>
        <kwd>Natural Language Processing</kwd>
        <kwd>Magical Towns</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The rapid growth of travel-review platforms has made it possible to collect vast amounts of textual
data reflecting tourists’ perceptions and satisfaction[
        <xref ref-type="bibr" rid="ref1 ref2 ref3 ref4">1, 2, 3, 4</xref>
        ]. TripAdvisor reviews of Mexico’s
Magical Towns, towns designated for their cultural heritage and touristic value, contain insights that can
guide local stakeholders in enhancing visitor experiences and shaping promotional activities. However,
extracting these insights automatically is challenging, especially when dealing with Spanish-language
nuances and tourism-specific terminology.
      </p>
      <p>
        Recent initiatives such as REST-MEX have emphasized the importance of robust sentiment
evaluation frameworks for tourism texts in Spanish [
        <xref ref-type="bibr" rid="ref5 ref6 ref7">5, 6, 7</xref>
        ]. Within IberLEF 2025, the REST-MEX shared task
requires systems to assign each review a sentiment score from 1 (very negative) to 5 (very positive),
determine whether it describes a hotel, restaurant, or attraction, and identify the correct Magical Town
among 40 to 60 options [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ]. Addressing these subtasks involves dealing with ambiguous expressions
of sentiment, overlapping mentions of site categories, and diverse ways users refer to locations.
      </p>
      <p>We decided to approach this challenge as a learning experience, so we applied diferent NLP
techniques to each sub-task independently. First, we used fine-tuning techniques for sentiment polarity
and review type classification, leveraging their ability to capture more complex contextual patterns
within the reviews. Next, we used more traditional techniques like SVM, just for educational purposes
and to be able to compare their performance. In parallel, we employed prompting on large language
models to assign reviews to the 40 Magical Towns and polarity, trusting these models’ capacity to infer
geographic and contextual relationships from a handful of examples (few-shot). By evaluating each
technique in isolation, we could compare their strengths and weaknesses and then combine the best
results from each approach to produce the final classifications.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Materials and methods</title>
      <p>This section outlines the diferent techniques we employed during our participation in the shared task.
Our approach focused on exploring a range of methodologies, from traditional machine learning to
more recent advancements in language modeling. Specifically, we experimented with classical methods
such as Support Vector Machines (SVM), fine-tuning of transformer-based models like RoBERTa (a
BERT variant pre-trained for Spanish), and prompting with various large language models (LLMs).
Each technique was applied with the goal of understanding its strengths and limitations in the context
of sentiment analysis, site-type classification, and the assignment of Magical Towns.</p>
      <p>For this contest, we were provided with a dataset consisting of 208,051 TripAdvisor reviews written
in Spanish and covering 40 diferent tourist destinations categorized as Magical Towns. The dataset
was suficiently large and well-structured, allowing us to train and evaluate our models without the
need for artificial data augmentation. This availability of real-world user-generated content enabled
us to test various approaches under realistic and meaningful conditions.</p>
      <sec id="sec-2-1">
        <title>2.1. Dataset Analysis</title>
        <p>First, we made an analysis of the provided data set. In this section, we are going to describe the
general characteristics of the dataset used in the REST-MEX 2025 sentiment evaluation task for Mexican
Magical Towns, and provide visual insights from it.</p>
        <p>The distribution of words per review, as we can see in Fig. 1, shows that the majority of reviews are
relatively short, with a predominant range between 20 and 200 words; longer reviews are rare. For
sitetype distribution (see Fig. 2), the dataset comprises three categories: Attraction, Hotel, and Restaurant.
Restaurant reviews are the most abundant, this might indicate a higher frequency in reviewing dining
experiences. All three categories represent a significant volume, so that the data distribution for this
subtask is quite well distributed, influencing good results.</p>
        <p>Finally, the data set exhibits a clear imbalance in sentiment polarity (see Fig. 3). The very positive
class reviews has a dominance over all other classes, this is due to that visitors usually share experiences
when they are highly satisfied. This imbalanced distribution will be taken into account for balancing
through diferent techniques (such as oversampling, under-sampling) or specialized loss functions to
improve classification performance for less represented classes.
2.2. SVM
We explored the performance of classical machine learning methods by employing Support Vector
Machines (SVM) for both the sentiment polarity and type classification subtasks. For the implementation
of the SVM models, we followed a structured approach described as follows:</p>
        <sec id="sec-2-1-1">
          <title>2.2.1. Data Preprocessing</title>
          <p>The text data was preprocessed using standard techniques to clean and normalize the reviews,
including:
• Conversion to lowercase.
• Removal of punctuation, numerical characters, null, and duplicate values.
• Stopword removal, stemming, and lemmatization.</p>
          <p>• Tokenization.</p>
        </sec>
        <sec id="sec-2-1-2">
          <title>2.2.2. Feature Extraction</title>
          <p>We utilized the Term Frequency-Inverse Document Frequency (TF-IDF) vectorization method to
convert the processed text data into numerical features suitable for SVM classification. TF-IDF captures in
a eficient way the importance of words within the reviews, aiding the SVM model in diferentiating
between categories and sentiment levels.</p>
        </sec>
        <sec id="sec-2-1-3">
          <title>2.2.3. Model Training</title>
          <p>Two separate SVM classifiers were developed: one for Type Classification (Categories: Attractive,
Hotel, Restaurant), and another for Sentiment Polarity Classification (Ratings from 1 to 5). The models
were trained using scikit-learn’s SVC implementation with linear kernels, which provided optimal
performance given the size and characteristics of the dataset.
2.2.4. Results
The dataset for the type classification task was relatively balanced, enabling the SVM model to achieve
robust performance. Evaluation metrics from the test set (313 samples) are shown in Table 1.</p>
          <p>The confusion matrix for this task is:
⎡60</p>
          <p>The model performed best at identifying hotel-related reviews, followed closely by restaurants, and
exhibited minor confusion distinguishing attractions.</p>
          <p>For the polarity classification task, we employed similar preprocessing and TF-IDF features.
However, this task proved more challenging due to the subjectivity and subtlety of sentiment expressions.
Overall performance was moderate, as shown below:
• Accuracy: 45%
• Macro Average F1-Score: 0.45
Class 1 (Very Negative): Precision = 0.53, Recall = 0.57, F1-Score = 0.55</p>
          <p>Higher performance suggests that strongly negative sentiments, expressed with clear cues, were
easier to detect.</p>
          <p>Class 2 (Negative): Precision = 0.36, Recall = 0.35, F1-Score = 0.35
Class 3 (Neutral): Precision = 0.36, Recall = 0.35, F1-Score = 0.36
Class 4 (Positive): Precision = 0.40, Recall = 0.39, F1-Score = 0.40
Class 5 (Very Positive): Precision = 0.59, Recall = 0.59, F1-Score = 0.59</p>
          <p>Highest performance among all classes, likely due to explicit positive expressions.</p>
          <p>Misclassifications were most common between adjacent classes—especially between Classes 3
(Neutral) and 4 (Positive), and between Classes 1 (Very Negative) and 2 (Negative). These patterns emphasize
the limitations of traditional SVM classifiers, which rely on surface-level lexical features and struggle
with subtler contextual distinctions.</p>
        </sec>
        <sec id="sec-2-1-4">
          <title>2.2.5. Discussion</title>
          <p>The SVM approach demonstrated strong baseline performance in structured tasks like type
classification, aided by clear lexical cues and a balanced dataset. In contrast, the sentiment polarity task revealed
the limitations of classical methods when dealing with more nuanced language.</p>
          <p>While SVM was successful in recognizing strongly negative and strongly positive sentiments, it
struggled with intermediate categories. Therefore, in the following sections, we explore
transformerbased fine-tuning and prompting methods, which are better suited to capture deeper contextual
information and improve performance on sentiment classification.</p>
        </sec>
      </sec>
      <sec id="sec-2-2">
        <title>2.3. Fine-tuning</title>
        <p>
          For the sentiment polarity and site-type classification subtasks, we employed a fine-tuning approach
using the dccuchile/bert-base-spanish-wwm-cased transformer model, a BERT variant
pretrained specifically on large Spanish corpora [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ], [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]. This choice was motivated by its proven
efectiveness in capturing syntactic and semantic nuances in Spanish, which is critical for tourism-related
user reviews.
        </p>
        <p>The fine-tuning process involved adding a classification head on top of the pre-trained encoder to
predict one of five sentiment classes or one of three site types. The model was implemented using
Hugging Face’s transformers library and trained using PyTorch. To leverage available hardware
eficiently, we dynamically assigned training to GPU if available:
device = torch.device("cuda" if torch.cuda.is_available() else "cpu")</p>
        <p>We loaded the pre-trained model and set it to output five classes for polarity or three for site type,
depending on the subtask, we run the model with 154227 instances, the maximum value we were
allowed to run in Colab. Importantly, we froze the first six layers of the BERT encoder to reduce
training time and overfitting risk, as the early layers primarily capture general linguistic information
that does not need to be adjusted for our specific classification tasks:
for name, param in model.bert.encoder.layer[:6].named_parameters():
param.requires_grad = False</p>
        <p>We used the AdamW optimizer with a learning rate of 1 × 10− 5, a common setting for transformer
ifne-tuning, along with CrossEntropyLoss as the objective function. This combination provides stable
convergence in multi-class classification tasks:
optimizer = AdamW(model.parameters(), lr=1e-5)
criterion = nn.CrossEntropyLoss()</p>
        <p>Training was conducted over a maximum of 10 epochs with early stopping based on validation loss
to prevent overfitting. After each epoch, we evaluated model performance on a held-out validation set,
computing both loss and accuracy. The best model (i.e., the one with the lowest validation loss) was
saved for future use:
model.save_pretrained(best_model_dir)
tokenizer.save_pretrained(best_model_dir)</p>
        <p>In addition to saving the model, we plotted loss and accuracy curves across epochs to visually inspect
convergence and learning behavior. These visualizations confirmed that our model improved steadily
during the first few epochs to around epoch 2, after,see figure 9 .</p>
        <p>After training, the model was evaluated on a separate test set using standard metrics: accuracy,
precision, recall, and F1-score. We also generated detailed classification reports for each class, allowing
us to analyze performance per label and detect any potential imbalance-related issues.</p>
        <p>The fine-tuning strategy proved to be especially efective for polarity classification, yielding
significantly higher accuracy than the SVM baseline. For site-type classification, the model also performed
well, achieving macro-F1 scores close to 0.96 in the oficial evaluation (see Table 4). However, as we will
discuss in later sections, this approach was not scalable for the 40-class Magical Town classification
subtask due to computational limitations.</p>
        <p>In summary, fine-tuning a Spanish BERT variant ofered a strong balance of performance and
practicality for sentiment and type classification subtasks. The combination of frozen layers, early
stopping, and appropriate hyperparameter choices helped us train an efective model without extensive
resources.</p>
      </sec>
      <sec id="sec-2-3">
        <title>2.4. Prompting</title>
        <p>Regarding prompting, we decided early on to apply it to the Magical Town classification subtask. This
decision was motivated by the high complexity of the task, as it involves assigning each review to one
of many possible classes. Although we had access to a large training dataset and it was theoretically
possible to solve this subtask properly by fine-tuning a transformer model such as BERT in Spanish,
we concluded that training a custom fine-tuned model for this multi-class problem was not feasible
due to the computational constraints of our environment.</p>
        <p>Instead, we opted for large language models and focused our eforts on prompting strategies.
Specifically, we used Google’s Gemini model through its API and conducted extensive testing with several
versions, both with and without reasoning capabilities. Our experiments aimed to find the right
balance between response quality, execution speed, and cost. For the final submission, we selected the
2.5-flash model, which was still in preview at the time.</p>
        <p>When implementing the prompting strategy, we chose not to perform any data preprocessing. Since
large language models (LLMs) are designed to handle raw, unstructured text efectively, applying
traditional preprocessing techniques such as lowercasing, stemming, or stopword removal is neither
necessary nor particularly beneficial in this context.</p>
        <p>
          To enhance the model’s performance in classification, we employed a few-shot prompting strategy,
where the model is guided by multiple input-output examples. This approach has been widely
recognized as efective for enabling language models to perform tasks without additional training, especially
in settings where labeled data is scarce or fine-tuning is not practical [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ].
        </p>
        <p>In addition, we explored both fine-tuning and prompting methods for the sentiment polarity
subtask, aiming to improve the initial results. For fine-tuning, we used a BERT based transformer model,
freezing the initial layers and training it for a 5-class classification task. This setup allowed us to
directly compare the two paradigms: prompting and fine-tuning. During implementation and evaluation,
prompting initially yielded better results; however, as we will show later in this paper, this advantage
didn’t hold on the test dataset.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Results</title>
      <p>Two diferent pipeline configurations were used at the time of submitting the results. In the first
configuration, we applied fine-tuning for both polarity classification and review type prediction, while
prompting was used for the Magical Town classification. In the second configuration, we replaced
the fine-tuned model for polarity classification with a prompting-based approach, keeping the other
components unchanged.</p>
      <p>
        The evaluation of the system follows the guidelines established in the Sentiment Analysis Track
provided by the contest organizers. The complete dataset for the task was divided by the organizers
into two subsets: one for training our models and the other for evaluating the results. The evaluation
process relies on standard metrics such as precision, recall, and F1-score, which are applied across
multiple subtasks, including polarity classification, type prediction (Attractive, Hotel, Restaurant), and
the identification of Magical Towns [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
      </p>
      <p>The oficial evaluation procedure defined by the organizers includes three main subtasks: polarity
classification, type prediction, and Magical Town identification. Each subtask is assessed using an
F-measure-based metric, and the final score is computed as a weighted average of the three.</p>
      <p>The polarity classification score is obtained by averaging the F-measure across all sentiment classes:
 () =
∑︀|=|1 ()
||</p>
      <p>For the type prediction subtask, which involves three categories (Attractive, Hotel, and Restaurant),
the macro-averaged F-measure is used:</p>
      <p>() +  () + ()
 () =
3</p>
      <p>The Magical Town identification subtask evaluates the system’s performance across all known towns
in a predefined list (MTL), averaging their F-measures:
(2)
 () =
∑︀=1( )   ()</p>
      <p>(  )</p>
      <p>Finally, the overall evaluation score combines the three subtasks, assigning higher importance to
polarity and Magical Town classification by applying weights of 2 and 3, respectively:
(1)
(3)</p>
      <p>As can be observed, the first run achieved marginally better results than the second, which is notably
surprising as it contrasts with the validation outcomes obtained during the model implementation and
prompting phases, where prompting demonstrated a slight advantage of 3% over our fine-tuned model.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusion</title>
      <p>In this work, we explored various natural language processing techniques for classifying
Spanishlanguage tourist reviews, applied to the REST-MEX 2025 challenge. The evaluated methods included
traditional approaches such as SVM, fine-tuning of Spanish-specialized transformer models
(DistilBERT), and prompting techniques (Gemini) with large language models.</p>
      <p>The results showed that traditional methods like SVM perform well on tasks with balanced classes
and clear lexical features, such as classifying the type of site (hotel, restaurant, attraction). However,
they perform poorly when classifying sentiment polarity.</p>
      <p>Fine-tuning BERT-based models for Spanish significantly improved performance, especially in
sentiment classification, thanks to their ability to understand complex contexts and linguistic nuances.
This approach also proved efective for site type classification, outperforming SVM in precision and F1
metrics. However, due to computational limitations, this method was not scalable for the 40-category
classification of Magical Towns.</p>
      <p>Finally, prompting with large language models proved to be a promising alternative for classifying
complex multi-class tasks, such as identifying the specific magical town, leveraging their few-shot
inference capabilities.</p>
      <p>Overall, combining the strengths of each technique can be an efective strategy to tackle complex
problems in the analysis of Spanish-language tourist reviews, maximizing both precision and eficiency
according to the specific subtask.</p>
    </sec>
    <sec id="sec-5">
      <title>Declaration on Generative AI</title>
      <p>We declare that the present manuscript has been written entirely by the authors and that no generative
artificial intelligence tools were used in its preparation, drafting, or editing.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Álvarez-Carmona</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-Gonzalez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Fajardo-Delgado</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. G.</given-names>
            <surname>Sánchez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Pérez-Espinosa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Martínez-Miranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bustio-Martínez</surname>
          </string-name>
          ,
          <article-title>Ángel DíazPacheco, Natural language processing applied to tourism research: A systematic review and future research directions</article-title>
          ,
          <source>Journal of King Saud University - Computer and Information Sciences</source>
          <volume>34</volume>
          (
          <year>2022</year>
          )
          <fpage>10125</fpage>
          -
          <lpage>10144</lpage>
          . URL: https://www.sciencedirect.com/science/article/pii/S1319157822003615. doi:https://doi.org/10.1016/j.jksuci.
          <year>2022</year>
          .
          <volume>10</volume>
          .010.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Álvarez-Carmona</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Aranda</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Guerrero-Rodríguez</surname>
            ,
            <given-names>A. Y.</given-names>
          </string-name>
          <string-name>
            <surname>Rodríguez-González</surname>
            ,
            <given-names>A. P.</given-names>
          </string-name>
          <string-name>
            <surname>López-Monroy</surname>
          </string-name>
          ,
          <article-title>A combination of sentiment analysis systems for the study of online travel reviews: Many heads are better than one</article-title>
          ,
          <source>Computación y Sistemas</source>
          <volume>26</volume>
          (
          <year>2022</year>
          ). doi:https: //doi.org/10.13053/CyS-26-2-4055.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Diaz-Pacheco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Álvarez-Carmona</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. A. C.</given-names>
            <surname>Chávez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Ramírez-Silva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <article-title>Artificial intelligence methods to support the research of destination image in tourism. a systematic review</article-title>
          ,
          <source>Journal of Experimental &amp; Theoretical Artificial Intelligence</source>
          <volume>0</volume>
          (
          <year>2022</year>
          )
          <fpage>1</fpage>
          -
          <lpage>31</lpage>
          . doi:
          <volume>10</volume>
          .1080/0952813X.
          <year>2022</year>
          .
          <volume>2153276</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodriguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Álvarez</surname>
          </string-name>
          <string-name>
            <surname>Carmona</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. P.</given-names>
            <surname>López-Monroy</surname>
          </string-name>
          ,
          <article-title>Studying online travel reviews related to tourist attractions using nlp methods: the case of guanajuato, mexico</article-title>
          ,
          <source>Current Issues in Tourism</source>
          <volume>26</volume>
          (
          <year>2023</year>
          )
          <fpage>289</fpage>
          -
          <lpage>304</lpage>
          . URL: https://doi.org/10.1080/13683500.
          <year>2021</year>
          .
          <volume>2007227</volume>
          . doi:
          <volume>10</volume>
          .1080/13683500.
          <year>2021</year>
          .
          <volume>2007227</volume>
          . arXiv:https://doi.org/10.1080/13683500.
          <year>2021</year>
          .
          <volume>2007227</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Álvarez-Carmona</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Aranda</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Arce-Cárdenas</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Fajardo-Delgado</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>GuerreroRodríguez</surname>
            ,
            <given-names>A. P.</given-names>
          </string-name>
          <string-name>
            <surname>López-Monroy</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Martínez-Miranda</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Pérez-Espinosa</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <article-title>Overview of rest-mex at iberlef 2021: Recommendation system for text mexican tourism</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>67</volume>
          (
          <year>2021</year>
          ). doi:https://doi.org/10.26342/2021-67-14.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <article-title>Álvarez-Carmona, Á</article-title>
          . Díaz-Pacheco,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <string-name>
            <surname>D. FajardoDelgado</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bustio-Martínez</surname>
          </string-name>
          ,
          <article-title>Overview of rest-mex at iberlef 2022: Recommendation system, sentiment analysis and covid semaphore prediction for mexican tourist texts</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>69</volume>
          (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <article-title>Álvarez-Carmona, Á</article-title>
          . Díaz-Pacheco,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-González</surname>
          </string-name>
          , L. BustioMartínez, V.
          <string-name>
            <surname>Muñis-Sánchez</surname>
            ,
            <given-names>A. P.</given-names>
          </string-name>
          <string-name>
            <surname>Pastor-López</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Sánchez-Vega</surname>
          </string-name>
          ,
          <article-title>Overview of rest-mex at iberlef 2023: Research on sentiment analysis task for mexican tourist texts</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>71</volume>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <article-title>Álvarez-Carmona, Á</article-title>
          . Díaz-Pacheco,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-González</surname>
          </string-name>
          , L. BustioMartínez, V.
          <article-title>Herrera-Semenets, Overview of rest-mex at iberlef 2025: Researching sentiment evaluation in text for mexican magical towns</article-title>
          , volume
          <volume>75</volume>
          ,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>González-Barba</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Chiruzzo</surname>
            ,
            <given-names>S. M.</given-names>
          </string-name>
          <string-name>
            <surname>Jiménez-Zafra</surname>
          </string-name>
          ,
          <article-title>Overview of IberLEF 2025: Natural Language Processing Challenges for Spanish and other Iberian Languages, in: Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2025), co-located with the 41st Conference of the Spanish Society for Natural Language Processing (SEPLN 2025), CEUR-WS</article-title>
          . org,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , Bert:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          ,
          <source>in: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>1</volume>
          (Long and Short Papers),
          <source>Association for Computational Linguistics</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          . URL: https://aclanthology.org/N19-1423.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>J. K.</given-names>
            <surname>Tripathy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. C.</given-names>
            <surname>Sethuraman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. V.</given-names>
            <surname>Cruz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Namburu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mangalraj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. N.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. S.</given-names>
            <surname>Ilango</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Vijayakumar</surname>
          </string-name>
          ,
          <article-title>Comprehensive analysis of embeddings and pre-training in nlp</article-title>
          ,
          <source>Computers and Electrical Engineering</source>
          <volume>93</volume>
          (
          <year>2021</year>
          )
          <article-title>107231</article-title>
          . doi:
          <volume>10</volume>
          .1016/j.compeleceng.
          <year>2021</year>
          .
          <volume>107231</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>S.</given-names>
            <surname>Ekin</surname>
          </string-name>
          ,
          <article-title>Prompt engineering for chatgpt: A quick guide to techniques, tips, and best practices</article-title>
          ,
          <source>TechRxiv</source>
          (
          <year>2023</year>
          ). URL: https://doi.org/10.36227/techrxiv.22683919.v1. doi:
          <volume>10</volume>
          .36227/techrxiv. 22683919.
          <year>v1</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>REST-MEX</surname>
            <given-names>2025</given-names>
          </string-name>
          ,
          <article-title>Rest-mex 2025: Data and evaluation, 2025</article-title>
          . URL: https://sites.google.com/cimat.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>