<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Prompt Engineering for Sentiment Analysis in Tourism: The Case of Mexican Pueblos Mágicos</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Federico Sandoval</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Algiedi Solutions, Cholula de Rivadavia</institution>
          ,
          <addr-line>Puebla</addr-line>
          ,
          <country country="MX">México</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>The rise of large language models (LLMs) has enabled new paradigms for performing natural language processing tasks without the need for fine-tuning. Among these paradigms, prompt engineering has emerged as a key technique for adapting generic models to specific domains. In this study, we explore the potential of prompt-based methods for sentiment analysis in the tourism sector, focusing on Spanish-language reviews of destinations within the “Pueblos Mágicos” program in Mexico. We design and evaluate a set of carefully crafted prompts using both zero-shot and few-shot settings, targeting various commercial LLMs. Our results show that efective prompt design can yield competitive performance for polarity classification without requiring extensive training, and that specific linguistic cues related to hospitality and culture significantly afect model behavior. This work ofers insights into the viability of prompt engineering for resource-constrained applications in domain-specific sentiment analysis, particularly in underrepresented languages like Spanish.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Prompt Engineering</kwd>
        <kwd>Sentiment Analysis</kwd>
        <kwd>Mexican Tourism</kwd>
        <kwd>Pueblos Mágicos</kwd>
        <kwd>NLP</kwd>
        <kwd>Large Language Models</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The exponential growth of user-generated content across digital platforms has radically transformed
how travelers share, assess, and access tourism experiences. Online reviews—ranging from brief social
media comments to detailed feedback on platforms like TripAdvisor or Google Reviews—have become
indispensable for understanding customer satisfaction, service quality, and destination reputation
[
        <xref ref-type="bibr" rid="ref1 ref2 ref3">1, 2, 3</xref>
        ]. These textual narratives ofer rich, fine-grained insight into the tourist journey, encompassing
emotional reactions, perceived authenticity, and even socio-political impressions of a place [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ]. In
Mexico, where tourism is one of the primary engines of economic development, harnessing such data is
critical for both public policy and private-sector competitiveness [
        <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
        ].
      </p>
      <p>
        Sentiment analysis, as a core task within natural language processing (NLP), has become a fundamental
tool in mining this content. It allows institutions and businesses to gauge the afective tone of public
opinion, thereby guiding service improvement, promotional strategies, and destination branding.
However, deploying sentiment analysis in practice—especially for Spanish-language content—poses
persistent challenges, such as limited resources, informal or noisy language, and the lack of annotated
data across domain-specific subregions [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>
        The Rest-Mex shared task, launched in 2021, has emerged as a benchmark for Spanish-language
sentiment analysis in tourism. Initially designed to evaluate models for satisfaction prediction and
polarity classification from tourist reviews [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], it has since evolved to encompass more complex tasks.
The 2022 edition, for example, included epidemic status classification from news articles [
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ], while
in 2023 the challenge expanded geographically to include Cuban and Colombian data, introducing
clustering-based tasks for opinion mining [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. Yet, across these editions, the backbone of the evaluation
has remained the ability to analyze sentiment and classify service types in multilingual and multicultural
tourist data.
      </p>
      <p>
        In the 2025 version of Rest-Mex, a new dimension was added by including geospatial granularity
via the identification of “Pueblos Mágicos”—a designation by the Mexican government for towns with
historical, cultural, or natural significance. Participants must now determine not only the sentiment
and service category of a review but also its geographic anchor among 40 possible towns [
        <xref ref-type="bibr" rid="ref13 ref14">13, 14</xref>
        ]. This
task amplifies the complexity of review classification, as it demands awareness of subtle contextual
cues that might imply a specific location, without being explicitly stated.
      </p>
      <p>
        Traditionally, addressing such challenges in NLP has involved pretraining and fine-tuning large
language models (LLMs) like BERT, RoBERTa, or BETO—the latter being a Spanish-specific adaptation
of BERT [
        <xref ref-type="bibr" rid="ref11 ref15">11, 15</xref>
        ]. Fine-tuning provides strong task performance, especially in resource-rich scenarios
where labeled data is abundant. However, this approach can be computationally intensive and inflexible
when rapid domain adaptation is required.
      </p>
      <p>Recently, a paradigm shift has emerged with the advent of *prompt engineering*, wherein LLMs are
guided using structured textual prompts rather than being retrained for each new task. This zero-shot
or few-shot prompting strategy leverages the knowledge already encoded in massive foundation models,
enabling them to generalize across tasks and domains with minimal adaptation. This shift is especially
promising for low-resource settings and underrepresented languages, such as Spanish in Latin American
contexts.</p>
      <p>Prompt engineering reframes traditional NLP challenges as controlled language generation problems.
Instead of optimizing parameters through gradient descent, the researcher optimizes the phrasing,
structure, and examples included in the prompt. When carefully crafted, prompts can elicit accurate
sentiment predictions, detect geographic indicators, and interpret informal or figurative expressions—all
without modifying the internal weights of the model. This makes prompt engineering particularly
suitable for domains like tourism, where language is diverse, creative, and rapidly evolving.</p>
      <p>In this paper, we explore how prompt engineering can be leveraged to perform sentiment analysis
on Spanish-language tourist reviews, with a specific focus on the “Pueblos Mágicos” of Mexico. We
construct and evaluate a set of prompts across various LLMs using zero-shot and few-shot techniques.
Our goal is to assess whether prompt-based sentiment classification can serve as a competitive alternative
to fine-tuned models in domain-specific, multilingual scenarios. By examining model outputs across
diferent prompt formulations, we aim to identify linguistic patterns that enhance performance and
gain insights into how large-scale models interpret cultural and contextual nuances within Mexican
tourism discourse.</p>
      <p>This work contributes to a growing body of research that seeks to democratize access to NLP
tools in low-resource languages and specialized domains. It also bridges methodological innovation
with practical application, ofering a framework for public agencies, tourism observatories, and small
businesses to derive actionable sentiment insights without requiring high-end infrastructure or deep
technical expertise.</p>
    </sec>
    <sec id="sec-2">
      <title>2. State of the Art</title>
      <p>The task of sentiment analysis in tourism intersects multiple domains of natural language processing,
particularly those concerned with afective computing, opinion mining, and socio-geographic text
understanding. In recent years, a substantial body of literature has explored how computational
techniques can extract emotional and evaluative signals from tourist narratives, typically through
supervised learning approaches on annotated corpora. Most early methods relied on bag-of-words or
lexical approaches, but the field has shifted toward deep learning models, especially Transformer-based
architectures, for their superior ability to model context, nuance, and sequential dependencies.</p>
      <p>
        In the tourism domain specifically, sentiment analysis has proven to be a valuable tool for capturing
visitor satisfaction, identifying pain points in services, and even detecting seasonality in perception
trends [
        <xref ref-type="bibr" rid="ref1 ref5">1, 5</xref>
        ]. However, the subjective nature of tourist reviews—often shaped by personal expectations,
cultural background, or transient experiences—makes them challenging to process reliably. These texts
are often informal, metaphorical, or hyperbolic, which further complicates rule-based or lexicon-driven
approaches.
      </p>
      <p>
        To address this, the community has embraced the use of pre-trained models fine-tuned for specific
sentiment classification tasks. One such model is BETO [ ? ], a Spanish-language version of BERT,
which has demonstrated competitive results in various tasks including sentiment analysis, named entity
recognition, and text classification [ 16]. Studies such as [
        <xref ref-type="bibr" rid="ref3">3, 17</xref>
        ] have successfully applied fine-tuned
BETO-based classifiers in tourism datasets, reporting substantial performance gains over traditional
baselines.
      </p>
      <p>
        The Rest-Mex shared task series has further fueled the development of Spanish NLP for tourism by
ofering large-scale, annotated corpora of reviews and structured challenges with multiple classification
tracks. Each year since 2021, Rest-Mex has introduced new complexities: from predicting satisfaction
scores and sentiment polarity, to classifying COVID-19 status from news [
        <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
        ], and even performing
thematic clustering across multiple countries [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]. While most top-performing systems rely on
finetuned Transformers, they also require significant infrastructure and domain adaptation, which may be
impractical in real-world deployment scenarios, particularly for small enterprises or local governments.
      </p>
      <p>This has given rise to growing interest in *prompt engineering* as a lightweight, scalable alternative
to fine-tuning. Prompt engineering refers to the art of crafting input queries that guide large language
models (LLMs) toward specific outputs, leveraging their pre-trained knowledge without altering their
weights. Models like GPT-3, GPT-4, PaLM, and LLaMA have demonstrated that prompt-based learning
can be efective across a wide range of tasks, including sentiment analysis, summarization, and question
answering [18, 19].</p>
      <p>In the context of sentiment analysis, prompt engineering enables users to frame polarity detection as
a natural language task. For example, instead of training a model to classify the sentence "The hotel was
amazing" as positive, a prompt-based system might be asked: “What is the sentiment of the following
review?” followed by the review text. The model, pre-trained on diverse data, can then generate a
response such as “positive” without the need for labeled training data.</p>
      <p>This zero-shot or few-shot paradigm is particularly attractive in Spanish and other under-resourced
languages, where annotated corpora are limited and task-specific models are scarce. Moreover, prompt
engineering allows rapid prototyping and domain testing by adjusting linguistic templates rather
than retraining full models. Techniques such as chain-of-thought prompting, few-shot examples, and
instruction tuning have further expanded the flexibility of this approach.</p>
      <p>Despite its promise, prompt engineering remains an emerging field with open questions about optimal
prompt design, robustness across domains, and cultural or linguistic biases embedded in LLMs. Some
studies have shown that prompt phrasing significantly afects outcomes, especially in subjective tasks
like sentiment analysis. Additionally, most existing benchmarks and evaluations remain concentrated
in English, creating an urgent need to assess prompt engineering in Spanish-language domains.</p>
      <p>In this work, we aim to bridge this gap by applying prompt engineering to a real-world
Spanishlanguage sentiment analysis task in tourism. Our focus on the "Pueblos Mágicos" dataset from Rest-Mex
2025 provides a rigorous and culturally grounded testbed for evaluating prompt design strategies in a
multilingual, domain-specific setting.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>Our methodological framework is grounded in the evaluation of prompt-based sentiment analysis
using large language models (LLMs), specifically within the context of Spanish-language tourism
reviews. Unlike prior approaches that involve fine-tuning model parameters, our study investigates
how carefully crafted prompts can direct the behavior of general-purpose LLMs to perform polarity
classification without explicit model training. We structure our methodology into three main stages:
dataset preparation, prompt design, and model evaluation.</p>
      <sec id="sec-3-1">
        <title>3.1. Dataset Description</title>
        <p>
          We conducted our experiments using the oficial training set provided by the Rest-Mex 2025 shared task
[
          <xref ref-type="bibr" rid="ref13 ref14">13, 14</xref>
          ], a large-scale benchmark dataset designed to advance sentiment and thematic analysis in the
tourism domain. The dataset consists of over 208,000 user-submitted reviews in Spanish, gathered from
prominent tourism platforms such as TripAdvisor, Booking.com, and Google Reviews.
        </p>
        <p>Each review in the dataset is annotated with three categorical labels: sentiment polarity (ranging
from 1 to 5), type of establishment (e.g., Hotel, Restaurant, Tourist Attraction), and a specific geographic
location among 40 designated Mexican towns classified as “Pueblos Mágicos.” These annotations support
multiple supervised learning tasks, but in our case, only the polarity label is used as the ground truth
reference for evaluating prompt performance.</p>
        <p>The reviews span a diverse linguistic landscape, reflecting the informal tone, regional expressions,
and varied formatting typical of user-generated content. Such variability presents a significant challenge
for automated analysis and provides an ideal setting to test the generalization capabilities of LLMs
through prompt-based interaction.</p>
        <p>To facilitate a robust evaluation, we filtered and cleaned the dataset to remove malformed or
incomplete records. In particular, we excluded entries missing the sentiment label or those with non-standard
characters likely to interfere with tokenization. The final corpus retained 207,689 reviews, each with
suficient length and semantic content to be meaningfully interpreted by a generative model.</p>
        <p>
          Figure 1 displays the distribution of sentiment classes across the dataset. As in previous editions of
Rest-Mex [
          <xref ref-type="bibr" rid="ref10 ref12">10, 12</xref>
          ], the data is imbalanced, with class 5 (most positive) being overrepresented, and class
1 (most negative) relatively rare. This imbalance underscores the need for evaluation metrics beyond
accuracy, such as macro-averaged F1-scores.
        </p>
        <p>
          While prior work has leveraged this dataset for fine-tuned classification using Transformers such
as BETO [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ], we treat it as a static evaluation corpus. No model is trained on the data. Instead, we
use the reviews as input queries to evaluate the response quality of pre-trained LLMs under diferent
prompting conditions.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Prompt Design and Strategies</title>
        <p>Prompt engineering requires careful consideration of linguistic structure, specificity, and clarity. In
our study, we crafted several prompt templates aimed at eliciting sentiment labels from the model. We
designed prompts to simulate typical human instructions, ranging from minimal (zero-shot) to more
guided (few-shot) forms. The prompt strategies fall into three categories:
• Zero-shot Instruction: The model is asked directly to classify sentiment from the review text,
using prompts like “Classify the sentiment of the following review on a scale from 1 (very negative)
to 5 (very positive):”
• Few-shot Examples: The prompt includes one to three labeled examples before the test input.</p>
        <p>This provides the model with demonstrations to infer the desired task structure and response
format.
• Contextual Cueing: Some prompts include background information about the “Pueblos Mágicos”
initiative or about tourism services, under the hypothesis that contextual framing may improve
the model’s interpretability of subtle cues.</p>
        <p>All prompts are evaluated using the same test samples from the cleaned Rest-Mex dataset. The
sentiment predicted by the model (typically a numerical output or a word token like “positive”) is
post-processed and mapped to the 1–5 polarity scale for metric comparison.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Model Selection and Query Protocol</title>
        <p>We tested our prompts across three major families of commercial and open-source LLMs, including:
• GPT-3.5 and GPT-4 (text-davinci-003, gpt-4) via the OpenAI API.
• LLaMA-based open models (e.g., Alpaca, Vicuna) deployed locally using HuggingFace pipelines.
• Google PaLM2 through the MakerSuite interface.</p>
        <p>All models were queried using a fixed protocol: same temperature (0.0 for deterministic outputs),
same prompt phrasing (for comparison across models), and a cap of 100 tokens per response. For
fairness, the review text was truncated at 500 tokens when necessary to remain within context window
limits.</p>
        <p>To reduce noise, each prompt-review pair was evaluated three times (where applicable), and the
modal predicted label was used for scoring. We also logged model confidence (when provided), latency,
and token usage to assess computational cost.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Evaluation Metrics</title>
        <p>The primary metric for model performance is the macro-averaged F1-score, which accounts for class
imbalance and reflects balanced precision-recall across sentiment levels. We also report overall accuracy
for comparison with existing fine-tuned models [ 17]. In addition, we performed qualitative error
analysis, focusing on reviews that yielded divergent predictions under diferent prompt types.</p>
        <p>Our methodology is designed to isolate the impact of prompt structure on model behavior, independent
of parameter tuning or domain-specific retraining. This allows us to evaluate the practical potential of
prompt engineering as a lightweight alternative for multilingual sentiment classification in real-world
tourism applications.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <p>The evaluation of our prompt engineering approach was conducted in two phases: (1) an internal
development phase using the training dataset, and (2) an oficial external evaluation phase using the
hidden test set provided by the organizers of Rest-Mex 2025. Given that our methodology did not
involve supervised fine-tuning, performance on the training set was expectedly modest, particularly for
sentiment polarity.</p>
      <sec id="sec-4-1">
        <title>4.1. Training Set Evaluation (Internal Validation)</title>
        <p>In the internal validation phase, prompts were applied to a stratified subset of 10,000 randomly selected
reviews from the training corpus. The macro F1-score for sentiment polarity was low across most
models, not exceeding 0.25. In particular, reviews with ambiguous or subtle emotional cues—such as
sarcastic comments or reviews with mixed opinions—posed significant dificulties for the models. The
few-shot prompting variant marginally improved stability but did not yield statistically significant gains
in accuracy or F1.</p>
        <p>This performance gap on the training set underscores the limitations of prompt-only strategies when
dealing with noisy, highly subjective, or domain-specific sentiment tasks in Spanish. It also confirms
the hypothesis that prompt engineering, while eficient, may require more specialized design to match
the efectiveness of supervised methods in low-resource settings.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Test Set Evaluation (Oficial Results)</title>
        <p>The oficial evaluation on the hidden Rest-Mex 2025 test set yielded a mixed but informative performance
profile. The results, summarized in Table 1, demonstrate that prompt engineering can provide meaningful
signals in certain subtasks, while still struggling in others—particularly when fine-grained diferentiation
is needed.</p>
        <sec id="sec-4-2-1">
          <title>Polarity</title>
        </sec>
        <sec id="sec-4-2-2">
          <title>Type</title>
          <p>For polarity classification, our prompt-based approach achieved a macro F1-score of 0.199 and an
accuracy of 13.4%. While this is significantly lower than scores typically obtained via fine-tuned models
(which often surpass 0.55 in F1), it is notable that the model was able to distinguish some degree of
polarity gradient without any parameter tuning. Precision and recall were highest for the most frequent
class (label 5), reflecting the class imbalance in the dataset.</p>
          <p>In contrast, performance on the type classification task was significantly better, with a macro F1-score of
0.333 and an accuracy of 82.5%. This result suggests that prompt-based reasoning is more efective when
the semantic distinctions between classes are clear and supported by consistent lexical patterns—such
as references to food, lodging, or attractions. These findings are aligned with previous studies indicating
that LLMs perform better in classification tasks with discrete and orthogonal categories [20].
The most challenging task by far was the identification of the correct “Pueblo Mágico” referenced in
each review. Here, the model achieved a macro F1-score of just 0.025 and a classification accuracy below
1%. Given the large number of classes (40) and the often implicit nature of geographic references in
the text, this result is not surprising. It highlights the dificulty of location grounding in LLMs without
access to external knowledge sources or contextual signals like geotagging.</p>
        </sec>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Error Patterns and Observations</title>
        <p>Qualitative analysis of the model’s responses revealed a few recurring trends. First, in the absence of
clear polarity markers, the models often defaulted to the majority class ("very positive"), which inflated
recall but suppressed precision. Second, the type classification task benefited from prompt templates
that explicitly named category examples ("e.g., Hotel, Restaurant, Attraction"), suggesting that concrete
framing aids LLM understanding. Third, the town task frequently failed due to hallucination—models
generated plausible but incorrect town names not present in the label space.</p>
        <p>Overall, while the results remain behind state-of-the-art supervised approaches, they validate the
core potential of prompt engineering as a flexible and lightweight alternative for certain subtasks. With
further refinement—such as prompt chaining, task-specific calibration, or use of external geographic
knowledge bases—performance in low-resource classification tasks could be substantially improved.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>This study explored the application of prompt engineering techniques for sentiment analysis in the
tourism domain, using Spanish-language reviews from the Rest-Mex 2025 dataset. By leveraging large
language models (LLMs) through zero-shot and few-shot prompting, we aimed to assess whether
such models could infer sentiment polarity and other attributes without the need for fine-tuning or
domain-specific retraining.</p>
      <p>Our results reveal a nuanced picture. While prompt-based methods yielded promising performance
in the classification of establishment type—achieving an accuracy above 82%—their efectiveness was
considerably lower in tasks requiring deeper contextual reasoning, such as sentiment polarity (F1macro
= 0.199) and especially town identification (F1 macro = 0.025). These outcomes suggest that LLMs, when
guided solely by prompts, can capture surface-level patterns and categorical cues, but struggle with
ifne-grained sentiment interpretation and geospatial grounding, particularly in resource-constrained
linguistic environments.</p>
      <p>Importantly, our findings validate the core potential of prompt engineering as a viable approach
for rapid prototyping, especially in low-resource settings where labeled data or computational power
is limited. Unlike traditional supervised methods, prompt-based approaches require no parameter
updates and can adapt quickly across domains. However, the tradeof lies in precision, control, and
interpretability.</p>
      <p>Future work should focus on improving prompt strategies through dynamic chaining, task-aware
calibration, and the integration of external knowledge sources such as tourism gazetteers or sentiment
lexicons. Additionally, a deeper exploration of linguistic and cultural factors in prompt design may help
unlock more consistent performance in subjective tasks like sentiment analysis.</p>
      <p>In sum, while prompt engineering is not yet a substitute for supervised fine-tuning in complex NLP
tasks, it represents a compelling and accessible frontier—especially for domain-specific applications
like tourism analytics in Spanish-speaking contexts.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>We declare that the present manuscript has been written entirely by the authors and that no generative
artificial intelligence tools were used in its preparation, drafting, or editing.
[16] I. Castillo-Ortiz, M. Á. Álvarez-Carmona, R. Aranda, Á. Díaz-Pacheco, Evaluating culinary skill
transfer: A deep learning approach to comparing student and chef dishes using image analysis,
International Journal of Gastronomy and Food Science 38 (2024) 101070.
[17] M. Á. Álvarez-Carmona, R. Aranda, R. Guerrero-Rodríguez, A. Y. Rodríguez-González, A. P.
LópezMonroy, A combination of sentiment analysis systems for the study of online travel reviews:
Many heads are better than one, Computación y Sistemas 26 (2022) 977–987.
[18] Y. Zhang, R. Yang, X. Xu, R. Li, J. Xiao, J. Shen, J. Han, Teleclass: Taxonomy enrichment and
llm-enhanced hierarchical text classification with minimal supervision, in: Proceedings of the
ACM on Web Conference 2025, 2025, pp. 2032–2042.
[19] M. Sallam, K. Al-Mahzoum, R. A. Almutawaa, J. A. Alhashash, R. A. Dashti, D. R. AlSafy, R. A.</p>
      <p>Almutairi, M. Barakat, The performance of openai chatgpt-4 and google gemini in virology
multiple-choice questions: a comparative analysis of english and arabic responses, BMC Research
Notes 17 (2024) 247.
[20] V. G. Morales-Murillo, H. Gómez-Adorno, D. Pinto, I. A. Cortés-Miranda, P. Delice, Lke-iimas
team at rest-mex 2023: Sentiment analysis on mexican tourism reviews using transformer-based
domain adaptation (2023).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodriguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Álvarez-Carmona</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Aranda</surname>
            ,
            <given-names>A. P.</given-names>
          </string-name>
          <string-name>
            <surname>López-Monroy</surname>
          </string-name>
          ,
          <article-title>Studying online travel reviews related to tourist attractions using nlp methods: the case of guanajuato, mexico</article-title>
          ,
          <source>Current issues in tourism 26</source>
          (
          <year>2023</year>
          )
          <fpage>289</fpage>
          -
          <lpage>304</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Diaz-Pacheco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Álvarez-Carmona</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Guerrero-Rodríguez</surname>
            ,
            <given-names>L. A. C.</given-names>
          </string-name>
          <string-name>
            <surname>Chávez</surname>
            ,
            <given-names>A. Y.</given-names>
          </string-name>
          <string-name>
            <surname>RodríguezGonzález</surname>
            ,
            <given-names>J. P.</given-names>
          </string-name>
          <string-name>
            <surname>Ramírez-Silva</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Aranda</surname>
          </string-name>
          ,
          <article-title>Artificial intelligence methods to support the research of destination image in tourism. a systematic review</article-title>
          ,
          <source>Journal of Experimental &amp; Theoretical Artificial Intelligence</source>
          <volume>36</volume>
          (
          <year>2024</year>
          )
          <fpage>1415</fpage>
          -
          <lpage>1445</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Álvarez-Carmona</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Aranda</surname>
            ,
            <given-names>A. Y.</given-names>
          </string-name>
          <string-name>
            <surname>Rodríguez-Gonzalez</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Fajardo-Delgado</surname>
            ,
            <given-names>M. G.</given-names>
          </string-name>
          <string-name>
            <surname>Sánchez</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Pérez-Espinosa</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Martínez-Miranda</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Guerrero-Rodríguez</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <article-title>Bustio-Martínez, Á. DíazPacheco, Natural language processing applied to tourism research: A systematic review and future research directions</article-title>
          ,
          <source>Journal of king Saud university-computer and information sciences 34</source>
          (
          <year>2022</year>
          )
          <fpage>10125</fpage>
          -
          <lpage>10144</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>E.</given-names>
            <surname>Olmos-Martínez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Álvarez-Carmona</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Aranda</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Díaz-Pacheco</surname>
          </string-name>
          ,
          <article-title>What does the media tell us about a destination? the cancun case, seen from the usa, canada, and mexico</article-title>
          ,
          <source>International Journal of Tourism Cities</source>
          <volume>10</volume>
          (
          <year>2024</year>
          )
          <fpage>639</fpage>
          -
          <lpage>661</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Á.</given-names>
            <surname>Díaz-Pacheco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Álvarez-Carmona</surname>
            ,
            <given-names>A. Y.</given-names>
          </string-name>
          <string-name>
            <surname>Rodríguez-González</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Aranda</surname>
          </string-name>
          ,
          <article-title>Quantifying diferences between ugc and dmo's image content on instagram using deep learning</article-title>
          ,
          <source>Information Technology &amp; Tourism</source>
          <volume>26</volume>
          (
          <year>2024</year>
          )
          <fpage>293</fpage>
          -
          <lpage>329</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>E. P.</given-names>
            <surname>Ramirez-Villaseñor</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Pérez-Espinosa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Álvarez-Carmona</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <article-title>Design, development, and evaluation of a chatbot for hospitality services assistance in spanish</article-title>
          ,
          <source>Acta universitaria 33</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S.</given-names>
            <surname>Arce-Cardenas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Fajardo-Delgado</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Álvarez-Carmona</surname>
            ,
            <given-names>J. P.</given-names>
          </string-name>
          <string-name>
            <surname>Ramírez-Silva</surname>
          </string-name>
          ,
          <article-title>A tourist recommendation system: a study case in mexico</article-title>
          ,
          <source>in: Mexican international conference on artificial intelligence</source>
          , Springer,
          <year>2021</year>
          , pp.
          <fpage>184</fpage>
          -
          <lpage>195</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Diaz-Pacheco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Álvarez-Carmona</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Carlos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <article-title>Measuring the diference between pictures from controlled and uncontrolled sources to promote a destination. a deep learning approach (</article-title>
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Álvarez-Carmona</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Aranda</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Arce-Cárdenas</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Fajardo-Delgado</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Guerrero-Rodríguez</surname>
            ,
            <given-names>A. P.</given-names>
          </string-name>
          <string-name>
            <surname>López-Monroy</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Martínez-Miranda</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Pérez-Espinosa</surname>
            ,
            <given-names>A. Y.</given-names>
          </string-name>
          <string-name>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <article-title>Overview of rest-mex at iberlef 2021: Recommendation system for text mexican tourism (</article-title>
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <article-title>Álvarez-Carmona, Á</article-title>
          . Díaz-Pacheco,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Fajardo-Delgado</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Guerrero-Rodríguez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bustio-Martínez</surname>
          </string-name>
          ,
          <article-title>Overview of rest-mex at iberlef 2022: Recommendation system, sentiment analysis and covid semaphore prediction for mexican tourist texts</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>69</volume>
          (
          <year>2022</year>
          )
          <fpage>289</fpage>
          -
          <lpage>299</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Alvarez-Carmona</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Aranda</surname>
          </string-name>
          ,
          <source>Determinación automática del color del semáforo mexicano del covid-19</source>
          a partir de las noticias (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <article-title>Álvarez-Carmona, Á</article-title>
          . Díaz-Pacheco,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Muñiz-Sánchez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. P.</given-names>
            <surname>López-Monroy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Sánchez-Vega</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bustio-Martínez</surname>
          </string-name>
          ,
          <article-title>Overview of rest-mex at iberlef 2023: Research on sentiment analysis task for mexican tourist texts</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>71</volume>
          (
          <year>2023</year>
          )
          <fpage>425</fpage>
          -
          <lpage>436</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <article-title>Álvarez-Carmona, Á</article-title>
          . Díaz-Pacheco,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Bustio-Martínez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Herrera-Semenets</surname>
          </string-name>
          ,
          <article-title>Overview of rest-mex at iberlef 2025: Researching sentiment evaluation in text for mexican magical towns</article-title>
          , volume
          <volume>75</volume>
          ,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>J.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>González-Barba</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Chiruzzo</surname>
            ,
            <given-names>S. M.</given-names>
          </string-name>
          <string-name>
            <surname>Jiménez-Zafra</surname>
          </string-name>
          ,
          <article-title>Overview of IberLEF 2025: Natural Language Processing Challenges for Spanish and other Iberian Languages, in: Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2025), co-located with the 41st Conference of the Spanish Society for Natural Language Processing (SEPLN 2025), CEUR-WS</article-title>
          . org,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Álvarez-Carmona</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Aranda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. Y.</given-names>
            <surname>Rodríguez-González</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Pellegrin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Carlos</surname>
          </string-name>
          ,
          <article-title>Classifying the mexican epidemiological semaphore colour from the covid-19 text spanish news</article-title>
          ,
          <source>Journal of Information Science</source>
          <volume>50</volume>
          (
          <year>2024</year>
          )
          <fpage>568</fpage>
          -
          <lpage>589</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>