<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A Multi-Task BETO-Based Framework with Synthetic Data Augmentation for Sentiment and Contextual Classification of Spanish Tourist Reviews</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Alvaro Zaid Gallardo-Hernández</string-name>
          <email>alvaro.gallardo@iberopuebla.mx</email>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ramón Aranda</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Angel Diaz-Pacheco</string-name>
          <email>angel.diaz@ugto.mx</email>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Centro de Investigación en Matemáticas</institution>
          ,
          <addr-line>Sede Mérida, Mérida, Yucatán</addr-line>
          ,
          <country country="MX">México</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Consejo Nacional de Humanidades, Ciencias y Tecnologías</institution>
          ,
          <addr-line>Ciudad de México</addr-line>
          ,
          <country country="MX">México</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Departamento de Ciencias e Ingenieras, Universidad Iberoamericana Puebla</institution>
          ,
          <addr-line>San Andrés Cholula, Puebla</addr-line>
          ,
          <country country="MX">México</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Departamento de Ingeniería Electrónica, División de Ingenierías, Universidad de Guanajuato - Campus Irapuato-Salamanca</institution>
          ,
          <addr-line>Yuriria</addr-line>
          ,
          <country country="MX">Mexico</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>IberLEF 2025</institution>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>This paper presents our solution for the Rest-Mex 2025 shared task, which involves multilingual sentiment and contextual classification of Spanish-language tourist reviews. Given a review, the task is to determine its sentiment polarity (from 1 to 5), the type of destination (hotel, restaurant, or attraction), and the corresponding Magical Town from a predefined list. To address this, we developed a multi-task classification model based on the BETO transformer, incorporating three output heads to predict polarity, type, and town simultaneously. To mitigate class imbalance-especially for underrepresented sentiment classes-we implemented a data augmentation strategy that combines Jaccard distance-based sampling with synonym substitution using the Spanish WordNet. This approach allowed us to synthetically generate diverse reviews for minority classes. The model was trained using PyTorch with the AdamW optimizer and evaluated using macro F1 and accuracy across all tasks. Our system achieved competitive results, particularly excelling in the opinion type classification subtask. Additionally, we employed generative AI tools such as Gemini 2.5 to assist in code generation and experimentation, highlighting the emerging role of LLMs in reproducible NLP research.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Sentiment Analysis</kwd>
        <kwd>Rest-Mex</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The rapid digitalization of the tourism industry has led to a proliferation of user-generated content,
particularly online reviews, which now play a central role in shaping traveler decisions and influencing
business strategies [
        <xref ref-type="bibr" rid="ref1 ref2 ref3">1, 2, 3</xref>
        ]. This paradigm shift has driven the need for automated systems capable of
extracting actionable insights from natural language data, especially in low-resource and domain-specific
contexts such as local tourism in Spanish-speaking regions [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        Unlike to others editions [
        <xref ref-type="bibr" rid="ref4 ref5 ref6">5, 6, 4</xref>
        ], the Rest-Mex 2025 [
        <xref ref-type="bibr" rid="ref7 ref8">7, 8</xref>
        ] shared task poses a multi-faceted
classification challenge that involves analyzing Spanish-language tourist reviews to predict three attributes:
sentiment polarity (ranging from 1 to 5), the type of destination (hotel, restaurant, or attraction), and
the specific town—typically drawn from a predefined list of Mexico’s Pueblos Mágicos [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Addressing
this task requires robust models that can handle multilingual input, unbalanced class distributions, and
nuanced domain-specific expressions.
      </p>
      <p>Recent advances in Natural Language Processing (NLP), particularly with the introduction of
transformer-based models such as BERT [9] and its Spanish counterpart BETO [10], have
significantly improved text classification performance in a wide range of tasks. BETO, trained specifically on
large Spanish corpora, has shown strong results in sentiment and domain adaptation tasks. Multi-task
learning strategies further enhance model generalization by allowing simultaneous optimization over
related objectives [11].</p>
      <p>However, data imbalance remains a persistent challenge in real-world scenarios. In the tourism
domain, positive reviews often dominate datasets, which can hinder model performance on minority
sentiment classes. To mitigate this, data augmentation techniques have been explored, including
lexical substitution using WordNet [12] and review recombination strategies based on lexical similarity
measures such as Jaccard distance [13]. These techniques aim to synthetically increase diversity in
underrepresented categories without compromising semantic plausibility.</p>
      <p>Furthermore, the integration of large language models (LLMs) into research workflows is reshaping
the NLP development landscape. Tools such as Gemini [14], ChatGPT, and Claude have shown potential
not only for code generation but also for experiment design and rapid prototyping [15]. As such, their
use introduces new opportunities for accelerating research while raising important questions about
reproducibility and human oversight.</p>
      <p>In this paper, we present a multi-task classification framework based on BETO for the Rest-Mex
2025 challenge. Our contributions include: (1) a synthetic data generation strategy combining Jaccard
sampling and WordNet-based synonym substitution, (2) a multi-head neural architecture leveraging
shared representations, and (3) a discussion on the use of LLMs to support NLP research. We evaluate
our approach using oficial task metrics and report competitive results, particularly in opinion type
classification.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Task Description: Sentiment Analysis</title>
      <p>The goal of this task is to analyze TripAdvisor reviews and classify them based on three main aspects:
sentiment polarity, type of site, and associated Pueblo Mágico (Magical Town).</p>
      <p>Each review provides useful insights into a tourist’s experience. First, the task involves predicting
the sentiment polarity of the review, assigning a score from 1 (very negative) to 5 (very positive) based
on the original rating given by the user.</p>
      <p>Next, participants must classify the type of site being reviewed. A review can refer to a hotel,
restaurant, or attraction, based on context and metadata.</p>
      <p>Finally, the third sub-task is to predict the Pueblo Mágico (Magical Town) associated with each review.
This classification is based on location metadata and aims to ensure that the review is accurately linked
to the correct town.</p>
      <sec id="sec-2-1">
        <title>Evaluation:</title>
        <p>Systems are evaluated using standard metrics such as precision, recall, and F1-score. The evaluation
is divided into three sub-tasks:
1. Polarity classification: The macro-F1 score over all polarity classes (1–5) is computed as:
 () =
∑︀|=|1 ()
||
where () is the F1-score for class  by system , and  = {1, 2, 3, 4, 5}.
2. Type classification: The macro-average F1-score over three categories (Attraction, Hotel,
Restaurant):</p>
        <p>() +  () + ()
 () =
3
where () is the F1-score for the Attraction class,  () for Hotel, and () for Restaurant.
3. Pueblo Mágico classification: A macro-F1 score is calculated over all towns in the Magical
Towns list (  ):
 () =
∑︀|=1 |   ()
|  |
where   () is the F1-score for the -th Magical Town.</p>
      </sec>
      <sec id="sec-2-2">
        <title>Final Score:</title>
        <p>The final score for a system  is computed as:
() = 2 ·  () +  () + 3 ·  ()
6</p>
        <p>This formula gives more weight to the polarity and Pueblo Mágico sub-tasks, reflecting their greater
importance in the evaluation.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <sec id="sec-3-1">
        <title>3.1. Data Preparation and Advanced Text Cleaning</title>
        <p>The original dataset, comprising Spanish-language reviews of Mexican tourist destinations, underwent
rigorous preprocessing to ensure data quality and consistency. This included handling missing values
(replacing NaN with empty strings), Unicode normalization, emoji removal, conversion to lowercase, and
substitution of URLs, user mentions, and hashtags with standardized placeholders (_URL_, _MENCION_,
_HASHTAG_). Extraneous punctuation and redundant whitespace were removed. Linguistic processing
with spaCy was applied for tokenization and lemmatization, as well as stopword and number removal.
Generic tags were preserved during processing.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Synthetic Data Generation and Class Balancing</title>
        <p>To address the pronounced class imbalance in sentiment polarity, we implemented a mixed strategy
based on both downsampling and synthetic oversampling.</p>
        <p>Synthetic Review Generation. For minority classes, we designed an algorithmic pipeline to create
lexically diverse and semantically plausible synthetic reviews, which operates as follows:
1. Selection of Base and Candidate Reviews: For each synthetic sample, a base review 1 is
randomly selected from the minority class. Ten additional candidate reviews {,} from the same
class are sampled at random.
2. Lexical Similarity via Jaccard Distance: For each candidate, we compute the Jaccard distance
between the base review and the candidate as follows:
 (, ) = 1 −
| ∩ |
| ∪ |
where  and  are the sets of alphabetic tokens (after lowercasing and stopword removal) from
the two reviews. This metric quantifies the lexical dissimilarity between reviews.
3. Candidate Selection: Based on the computed distances, three key candidates are identified: the
most similar (), the most distant (), and an intermediate candidate ().
4. Fragment Extraction and Synonym Substitution: From 1, , , and , a random subset
(typically 25%) of their tokens is selected. To further diversify the generated text, each selected
token is optionally replaced with a synonym using the Spanish WordNet corpus (via NLTK and
OMW 1.4). If a suitable synonym is available and distinct from the original token, it is used;
otherwise, the original token is retained.
5. Synthetic Review Assembly: The selected fragments (enriched with synonyms) are
concatenated to synthesize a new artificial review, which inherits the label of the target class.</p>
        <p>This process is repeated iteratively until the minority class reaches the desired number of samples.
Class Balancing. To address the pronounced class imbalance across all three classification tasks, we
implemented a generalized balancing strategy involving both random downsampling and synthetic
oversampling. This procedure was applied independently to the sets of labels for polarity, destination
type, and Magical Town, with the aim of creating a more uniform data distribution for the model.</p>
        <p>Let  be the set of classes for a given task  ∈ {polarity, type, town}. Our objective is to transform
the original sample distribution, where each class  ∈  has a size of ||, into a balanced distribution
with a new size |′|. This is achieved by defining a target sample size, target, and applying the following
transformation:
|′| =
⎧⎪target if || &gt; target (Downsampling)
⎨</p>
        <p>target if || &lt; target (Oversampling)
⎪⎩||</p>
        <p>if || ≈ target (Invariant)
where downsampling is performed by random selection without replacement, and oversampling is
achieved using the synthetic review generation method described previously.</p>
        <p>For the polarity task, this strategy was applied with a target size of target = 15519. The majority
classes (polarities 4.0 and 5.0) were downsampled, while the minority classes (1.0 and 2.0) were
synthetically oversampled to reach target. The neutral class (polarity 3.0), whose size was close to the target,
was kept invariant, preserving its original data.</p>
        <p>A similar principle was applied to the destination type and Magical Town tasks. For each, we
identified majority and minority classes relative to a target size and applied downsampling or synthetic
oversampling, respectively, to mitigate the skew in their distributions.</p>
        <p>The final comprehensive dataset, composed of a mix of original, downsampled, and synthetic reviews
for all tasks, is then shufled and prepared for model training.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Data Preparation for PyTorch</title>
        <p>The balanced dataset is further processed for compatibility with PyTorch. Categorical textual labels
(e.g., sentiment polarity, destination type, and town) are mapped to integer values via label encoding.
Each review is tokenized using the pre-trained BETO tokenizer, with sequences padded or truncated to
a fixed maximum length max. The dataset is then stratified into training and validation sets. PyTorch
DataLoader objects are instantiated to manage mini-batch sampling and data shufling.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Neural Network Architecture: MultiTaskBETO</title>
        <p>For classification, we employ a multi-task neural network architecture based on the BETO transformer
model, specialized for Spanish. The architecture includes:
• BETO Encoder: The core of the model is the pre-trained BETO encoder
(AutoModel.from_pretrained), which generates contextualized vector representations for
each input review.
• Dropout Regularization: A dropout layer is applied to the pooled output to reduce overfitting.
• Multi-task Classification Heads: Three independent linear layers receive the pooled output
(corresponding to the [CLS] token):
1. Sentiment polarity (5 classes)
2. Destination type (hotel, restaurant, attraction)
3. Magical Town (up to 60 classes)</p>
        <p>Each head outputs unnormalized logits for its corresponding task.
3.5. Training Pipeline
The model is trained using the AdamW optimizer and categorical cross-entropy loss for each
classiifcation head. Let ℒpol, ℒtype, and ℒtown denote the losses for sentiment, type, and town prediction,
respectively; the total loss for each batch is computed as:</p>
        <p>ℒtotal = ℒpol + ℒtype + ℒtown
The model is trained for  epochs, where in each epoch the model iterates through all training batches,
computes forward and backward passes, updates model weights, and evaluates performance on the
validation set.</p>
      </sec>
      <sec id="sec-3-5">
        <title>3.6. Iterative Training and Evaluation</title>
        <p>For each epoch, training and validation loss, as well as macro-F1 and accuracy for each task, are
monitored and recorded. Upon completion, the best model (according to validation performance) is
selected for further analysis and reporting.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <p>We obtained an overall Track Score of 0.6145. In particular, our system excelled in the opinion type
classification subtask, where we achieved a Macro F1 score of 0.9688. This indicates that our model
was highly efective in diferentiating between reviews of hotels, restaurants, and attractions.</p>
      <p>For the polarity classification task, we reached a classification accuracy of 71.52% , which was
among the highest in the evaluation. However, the Macro F1 score for this subtask was lower, at 0.4381,
highlighting performance imbalances across sentiment classes. Specifically, our system performed best
on class 4 with an F1-score of 0.8525, while performance was notably weaker on minority classes
such as class 1, which yielded an F1-score of 0.1856.</p>
      <p>In the town classification task, we obtained a Macro F1 score of 0.5902. Although slightly below
the scores achieved by the top-performing systems, this result still reflects a reasonable performance
given the large number of distinct town labels in the dataset.</p>
      <p>Our system demonstrated strong capabilities in the opinion type classification task and achieved
high accuracy in polarity classification. Nonetheless, the results suggest that future work should focus
on improving the balance of predictions across all sentiment classes and enhancing performance on the
more granular town classification subtask.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusions</title>
      <p>In this paper, we presented a multi-task classification approach for the Rest-Mex 2025 shared task,
focusing on the automatic prediction of sentiment polarity, opinion type, and associated town in
Spanish-language tourist reviews. Our system was built upon the BETO transformer and demonstrated
competitive performance—particularly in the opinion type classification task, where it achieved
nearperfect macro F1 results.</p>
      <p>To mitigate the strong class imbalance present in the original dataset, we applied a synthetic data
generation pipeline using Jaccard-based sampling and synonym substitution via Spanish WordNet. This
approach increased lexical diversity and helped improve performance in minority classes, although
challenges remained—especially in sentiment classes with few training examples.</p>
      <p>For future work, we intend to explore and compare alternative architectures, including
encoderdecoder models and large language models (LLMs) such as LLaMA, Mistral, and other multilingual
transformers. We also recognize the increasing relevance of generative AI tools in research workflows.
While this work leveraged Google’s Gemini 2.5 Pro for code generation and experiment support,
further experimentation is needed to systematically compare the contributions of various LLM-based
assistants—such as ChatGPT, Claude, or open-source alternatives—on reproducibility, model design,
and code quality.</p>
      <p>By incorporating these tools into a reproducible benchmarking framework, we aim to assess not only
model performance but also the qualitative impact of AI-assisted research, ultimately enhancing the
efectiveness and transparency of sentiment analysis in multilingual tourism datasets.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements</title>
      <p>The authors gratefully acknowledge the support provided by the Mexican Academy of Tourism Research
(AMIT) for the project “Balancing Tourism Text Data with Artificial Intelligence for Sentiment Analysis: A
Specialized Language Model Approach” funded through the Research Projects 2024 call. Additionally, this
work was also supported by the project “Text Generation for Data Balancing in Sentiment Classification:
Application to Tourism Data” under the CICIMPI 2024 call of the Centro de Investigación en Matemáticas
(CIMAT).</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>We declare that the present manuscript has been written entirely by the authors and that no generative
artificial intelligence tools were used in its preparation, drafting, or editing.
[9] J. Devlin, M.-W. Chang, K. Lee, K. Toutanova, Bert: Pre-training of deep bidirectional transformers
for language understanding, in: Proceedings of NAACL-HLT, 2019.
[10] J. Cañete, G. Chaperon, R. Fuentes, J. Pérez, D. Rogers, Spanish pre-trained bert model and
evaluation data, 2020. ArXiv:2003.12171.
[11] S. Ruder, An overview of multi-task learning in deep neural networks, arXiv preprint
arXiv:1706.05098 (2017).
[12] G. A. Miller, Wordnet: A lexical database for english, Communications of the ACM 38 (1995)
39–41.
[13] P.-N. Tan, M. Steinbach, V. Kumar, Introduction to Data Mining, Pearson, 2005.
[14] G. DeepMind, Gemini 1.5 technical report, 2024. Https://deepmind.google/technologies/gemini/.
[15] OpenAI, Gpt-4 technical report, 2023. Https://openai.com/research/gpt-4.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>F.</given-names>
            <surname>Calderón</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Blanco</surname>
          </string-name>
          , Impacto de internet en el sector turístico,
          <source>Revista UNIANDES Episteme</source>
          <volume>4</volume>
          (
          <year>2017</year>
          )
          <fpage>477</fpage>
          -
          <lpage>490</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>E.</given-names>
            <surname>Olmos-Martínez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Álvarez-Carmona</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Aranda</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Díaz-Pacheco</surname>
          </string-name>
          ,
          <article-title>What does the media tell us about a destination? the cancun case, seen from the usa, canada, and mexico</article-title>
          ,
          <source>International Journal of Tourism Cities</source>
          <volume>10</volume>
          (
          <year>2023</year>
          )
          <fpage>639</fpage>
          -
          <lpage>661</lpage>
          . URL: http://dx.doi.org/10.1108/IJTC-09-2022-0223. doi:
          <volume>10</volume>
          .1108/ijtc-09-2022-0223.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Álvarez-Carmona</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          , et al.,
          <article-title>Big data analytics of online news to explore destination image using a comprehensive deep-learning approach: a case from mexico</article-title>
          ,
          <source>Information Technology &amp; Tourism</source>
          <volume>26</volume>
          (
          <year>2024</year>
          )
          <fpage>147</fpage>
          -
          <lpage>182</lpage>
          . URL: https://doi.org/10.1007/ s40558-023-00278-5. doi:
          <volume>10</volume>
          .1007/s40558-023-00278-5.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Alvarez-Carmona</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Díaz-Pacheco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          , et al.,
          <article-title>Overview of rest-mex at iberlef 2023: Research on sentiment analysis task for mexican tourist texts</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>71</volume>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Álvarez-Carmona</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Aranda</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Arce-Cárdenas</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Fajardo-Delgado</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Guerrero-Rodríguez</surname>
            ,
            <given-names>A. P.</given-names>
          </string-name>
          <string-name>
            <surname>López-Monroy</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Martínez-Miranda</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Pérez-Espinosa</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <article-title>Overview of rest-mex at iberlef 2021: Recommendation system for text mexican tourism</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>67</volume>
          (
          <year>2021</year>
          ). doi:https://doi.org/10.26342/2021-67-14.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <article-title>Álvarez-Carmona, Á</article-title>
          . Díaz-Pacheco,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Fajardo-Delgado</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bustio-Martínez</surname>
          </string-name>
          ,
          <article-title>Overview of rest-mex at iberlef 2022: Recommendation system, sentiment analysis and covid semaphore prediction for mexican tourist texts</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>69</volume>
          (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <article-title>Álvarez-Carmona, Á</article-title>
          . Díaz-Pacheco,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bustio-Martínez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Herrera-Semenets</surname>
          </string-name>
          ,
          <article-title>Overview of rest-mex at iberlef 2025: Researching sentiment evaluation in text for mexican magical towns</article-title>
          , volume
          <volume>75</volume>
          ,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>González-Barba</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Chiruzzo</surname>
            ,
            <given-names>S. M.</given-names>
          </string-name>
          <string-name>
            <surname>Jiménez-Zafra</surname>
          </string-name>
          ,
          <article-title>Overview of IberLEF 2025: Natural Language Processing Challenges for Spanish and other Iberian Languages, in: Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2025), co-located with the 41st Conference of the Spanish Society for Natural Language Processing (SEPLN 2025), CEUR-WS</article-title>
          . org,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>