<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>LPQ Team at HOPE 2025: Multilingual Hope Speech Detection Using BERTology and LLM Adaptation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Le Phu Quy</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Dang Van Thin</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Information Technology-VNUHCM</institution>
          ,
          <addr-line>Quarter 6, Linh Trung Ward, Thu Duc District, Ho Chi Minh City</addr-line>
          ,
          <country country="VN">Vietnam</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Vietnam National University</institution>
          ,
          <addr-line>Ho Chi Minh City</addr-line>
          ,
          <country country="VN">Vietnam</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>This paper presents our approach for detecting hope speech in social media posts written in English and Spanish. The system employs integrated transformer models, while separate experiments with the LLaMA model are conducted for emotion detection. Based on the PolyHope task at IberLEF 2025, the framework supports both binary classification-separating hope speech from non-hope speech-and a detailed multiclass categorization that distinguishes among generalized, realistic, and unrealistic expressions of hope. Experimental results show strong performance across diferent models and languages.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Hope classification</kwd>
        <kwd>Spanish language</kwd>
        <kwd>English language</kwd>
        <kwd>sentiment analysis</kwd>
        <kwd>fine-tuning BERT</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Hope is an essential human emotion that shapes behavior, mood, and decision-making across diverse
cultural contexts, and it fills our lives with both resilience and aspiration. In today’s digital era, social
media platforms have become vibrant arenas where people share intimate personal dreams alongside
collective visions for a brighter future. While the academic challenge of detecting and analyzing
hope speech is made more complex by its subtle blend of irony, sarcasm, and multilingual variations,
developing robust methodologies to capture these nuances remains crucial. Not only does such work
contribute to our scholarly understanding of emotional expression, but it also supports practical eforts
in mental health monitoring, crisis intervention, and designing public policies that truly resonate with
human experience.</p>
      <p>
        The PolyHope shared task at IberLEF 2025 [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] aims to detect and classify hope speech in social media
posts written in English and Spanish. This efort is significant in the field of computing and sentiment
analysis because it attempts to diferentiate various forms of hope instead of limiting captures as a
mere positive or negative classification. The two main tasks in this challenge are: (1) hope speech
detection using binary classification which involves hopeful messages and non-hopeful messages, and
(2) multiclass diferentiation of general, realistic, and unrealistic hope alongside sarcastic remarks.
      </p>
      <p>In this work, we present our system for detecting hope speech in social media posts across both
English and Spanish. By integrating leading transformer models—DeBERTa and DistilBERT for English,
and BETO alongside RoBERTa-BNE for Spanish—while also using LLaMA for emotion detection, our
system addresses both binary and multiclass classification (capturing generalized, realistic, unrealistic,
and sarcastic forms of hope). Extensive experiments on the PolyHope 2025 dataset not only confirm the
robustness of our approach but also pave the way for practical applications in mental health monitoring,
crisis response, and political discourse analysis.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Transformer-based models have greatly changed how we handle many NLP tasks, including emotion and
sentiment analysis. For example, BERT [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and its variants perform very well on sentiment benchmarks.
For multilingual tasks, models like XLM-R [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] and mBERT [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] let us transfer learning from languages
with lots of data to those with less data. In hope speech detection, transformer approaches have also
shown promising results. At IberLEF 2023, the I2C-Huelva team [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] reached the top rankings by using
BERTuit for Spanish and BERT for English. Similarly, NLP URJC [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] used BERT for English and BETO
for Spanish. These cases support the trend that transformer-based models generally beat traditional
machine learning methods in emotion detection [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        Language-specific transformer models are very efective for emotion tasks. DeBERTa [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] uses a
special attention mechanism and an improved mask decoder, leading to state-of-the-art performance on
English sentiment analysis. For Spanish, BETO [9] and RoBERTa-BNE [10] are strong choices; notably,
RoBERTa-BNE was trained on a massive corpus of Spanish text from the National Library of Spain.
Large language models (LLMs) such as GPT-3 [11] and LLaMA [12] have boosted emotion and sentiment
analysis by understanding subtle language cues. For example, the Zootopi team at IberLEF 2023 used
prompt-based ChatGPT to detect hope speech, winning the Spanish subtask [13]. To reduce the high
computational cost of fine-tuning huge models, researchers now use Parameter-Eficient Fine-Tuning
(PEFT) methods like Low-Rank Adaptation (LoRA) [14].
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <sec id="sec-3-1">
        <title>3.1. Data Preprocessing</title>
        <p>In the early stages of our research, we experimented with a variety of preprocessing techniques—including
text normalization, comprehensive punctuation removal, and aggressive stop-word filtering—in an
efort to enhance our model’s performance in detecting hope speech. However, we found that these
approaches often eliminated important contextual and emotional nuances that are crucial for
understanding the subtleties of human communication. Based on our experimental insights, we ultimately
adopted a minimal preprocessing strategy. This approach relies on standard, model-specific tokenization
and essential sequence management, ensuring that the delicate linguistic details remain intact. Such
a strategy not only preserves the richness of the original text but also supports our aim of achieving
robust and sensitive emotion detection in both English and Spanish social media content.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. BERT-Base Classification</title>
        <p>Recent studies have demonstrated the efectiveness of BERT-based approaches for text classification
tasks. All models are fine-tuned within a unified framework to optimize performance on our hope
speech classification task. For the BERT-based models, we minimize a standard cross-entropy loss
using hyperparameters recommended by recent literature and validated through empirical studies.
This BERT-based classification framework efectively combines deep and contextualized language
representations with task-specific fine-tuning, making it well-suited for capturing subtle cues in text.
Its widespread adoption in recent research underscores its robustness and efectiveness across diverse
text classification settings.</p>
        <p>The following models were incorporated in our experiments:
• DeBERTa: Utilizes a disentangled attention mechanism to enhance contextual representation.
• DistilBERT: A lighter version of BERT, retaining approximately 97% of its performance while
significantly reducing computational overhead.
• BETO: A Spanish-specific adaptation of the BERT architecture, optimized to capture native
linguistic nuances.
• RoBERTa-BNE: A RoBERTa variant fine-tuned on extensive Spanish corpora, enabling efective
handling of regional dialects and stylistic variations.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. LLaMA Hybrid Approach</title>
        <p>Large language models perform well in many NLP tasks, yet adapting them for specialized domains
like hope speech classification requires balancing eficiency with domain-specific adjustments. The
LLaMA hybrid approach achieves this by combining parameter-eficient adaptation and structured
instruction tuning, preserving the model’s contextual depth while enhancing its sensitivity to subtle
signals of hope. The framework follows a four-stage process: Low-Rank Adaptation, structured prompt
engineering, targeted instruction tuning, and post-processing.</p>
        <p>1. LoRA Configuration: Low-Rank Adaptation (LoRA) is employed by decomposing weight
updates into low-rank matrices. This strategy enables fine-tuning of the LLaMA model by
adjusting only a small subset of its parameters, thereby preserving pre-trained performance while
optimizing computational eficiency. For these experiments, a rank of  = 16 and a scaling factor
 = 32 were selected, providing an optimal trade-of between adaptation capacity and resource
demands.
2. Structured Prompt Engineering: The task is rephrased as a guided conversation with clear
instructions to help the model detect hope. Detailed guidelines help the model identify clear
signs of hope, subtle hints, and the intended meaning behind sarcastic remarks. This organized
method ensures the model stays sensitive to the fine nuances of human communication.
3. Instruction Tuning: The SFTTrainer uses a response-focused loss function to increase the
likelihood of correct labels. This tuning method takes into account previously generated tokens and
the parameters adjusted by LoRA, fine-tuning the model for accurate hope speech classification
while keeping computations eficient.
4. LLaMA Post-Processing: A dedicated post-processing module extracts and refines predictions
from the generated text. Using multiple extraction strategies, it efectively handles variations in
output formats, ensuring consistent and reliable classification results.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments</title>
      <sec id="sec-4-1">
        <title>4.1. Dataset</title>
        <p>The dataset [15, 16, 17, 18, 19, 20, 21] consists of English (en) and Spanish (es) social media texts
annotated for two tasks: binary classification (Hope vs. Not Hope) and multiclass classification. Statistics
are summarized below:</p>
        <p>Not Hope</p>
        <p>Generalized Hope</p>
        <p>Realistic Hope</p>
        <p>Unrealistic Hope Sarcasm</p>
        <sec id="sec-4-1-1">
          <title>Language</title>
          <p>English
Spanish</p>
        </sec>
        <sec id="sec-4-1-2">
          <title>Split</title>
          <p>Train
Dev
Train
Dev</p>
        </sec>
        <sec id="sec-4-1-3">
          <title>Samples</title>
        </sec>
        <sec id="sec-4-1-4">
          <title>Total</title>
        </sec>
        <sec id="sec-4-1-5">
          <title>Not Hope (%) Hope (%)</title>
        </sec>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Experiment Setting</title>
        <p>The experimental setup aimed to assess model performance in classifying English and Spanish social
media texts expressing hope, across both binary and multiclass tasks. To ensure a balanced comparison,
four transformer-based architectures were tested. DeBERTa (microsoft/deberta-base) stood out for
its advanced attention mechanism, which improves understanding of complex language patterns.
DistilBERT (distilbert-base-uncased) ofered a faster, lightweight alternative to BERT, maintaining
97% of its performance with fewer computational demands. RoBERTa-BNE
(BSC-TeMU/roberta-basebne) was specialized for Spanish, leveraging data from the National Library of Spain to refine its
understanding of the language. Similarly, BETO (dccuchile/bert-base-spanish-wwm-cased) adapted
BERT’s framework for Spanish, using whole-word masking to better capture meaning.
Consistency in training was a priority across all experiments. Each model was fine-tuned with a
batch size of 8, using the AdamW optimizer at a learning rate of 2 × 10 −5 and a weight decay of 0.01.
Models were trained for 5 epochs with early stopping, and input text was limited to 128 tokens through
truncation or padding. These parameters adhered to best practices for transformer models while also
ensuring computational eficiency.</p>
        <p>For the Llama-3 experiments (unsloth/Meta-Llama-3.1-8B-bnb-4bit), adjustments were made
to address resource limitations. Techniques such as 4-bit quantization, implemented via bitsandbytes,
reduced GPU memory usage by 70%. Low-Rank Adaptation (LoRA) was applied with a rank of  = 16
and scaling factor  = 32 , allowing eficient fine-tuning by updating only 0.1% of the model’s parameters.
Prompts were tailored to the task: binary classification relied on simple instructions focused on detecting
hope versus no hope, while multiclass classification required more detailed prompts to distinguish
between nuanced categories, including generalized, realistic, unrealistic hope, and sarcasm. Training
for Llama-3 utilized an extended context window of 512 tokens, a higher learning rate of 2 × 10 −4 ,
and a shorter training cycle of 3 epochs to manage computational constraints. This experimental
design ensured a robust evaluation, comparing traditional transformer-based architectures against the
advanced capabilities of Llama-3 while maintaining eficiency and academic rigor</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Main Result</title>
        <p>DeBERTa achieves state-of-the-art performance in both binary (84.76% F1) and multiclass (75.83% F1)
hope speech detection, outperforming DistilBERT and Llama 3.1 8B across languages. While transformer
models generalize robustly, multiclass tasks remain challenging due to nuanced hope subtypes and
sarcasm. Llama 3.1 8B shows promise but lags in multiclass scenarios (72.56% F1), suggesting room for
improved fine-tuning strategies. Results validate the eficacy of task-specific architectures over LLM
approaches.</p>
        <p>Our approach demonstrates strong performance across a diverse range of classification tasks. In
particular, our model achieved first rank in Multiclass Spanish, recording a weighted F1 score of 0.7417,
and secured third place in Spanish Binary with an F1 score of 0.8446. These results highlight the
robustness of language-specific tuning in Spanish contexts. However, for English tasks, our model
performs worse on English tasks, ranking seventh in Binary and fourth in Multiclass. This lower
performance may be due to the greater complexity of the English dataset and less efective fine-tuning.
Detailed evaluation shows that transformer models like DeBERTa and BETO generally outperform the
Llama 3.1 8B setup by approximately 2–4% F1.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Error Analysis and Discussion</title>
      <p>Our analysis revealed two primary error patterns afecting model performance. First, the models
struggled to accurately distinguish between subtle hope subtypes, particularly in Spanish texts.
Differentiating between Realistic Hope and Unrealistic Hope proved especially dificult, as reflected in
F1-scores ranging from 0.51 to 0.56. Second, data imbalance played a significant role in these challenges.
The Sarcasm class accounts for only 6.2% of the Spanish training data and exhibited low recall, between
1.a English
1.b Spanish
2.a English
2.b Spanish</p>
      <p>DeBERTa
DistilBERT
Llama 3.1 8B
BETO
RoBERTa-BNE
Llama 3.1 8B
DeBERTa
DistilBERT
Llama 3.1 8B
BETO
RoBERTa-BNE</p>
      <p>Llama 3.1 8B
Task
English Binary
English Multiclass
Spanish Binary</p>
      <p>2
Spanish Multiclass 3
1
1
2
7
1
2
4
1
2
3
Subtask</p>
      <p>Model</p>
      <p>Accuracy</p>
      <p>Precision</p>
      <p>Recall</p>
      <p>F1 Score</p>
      <p>Binary Classification
0.60 and 0.77, because the models tended to favor the majority Not Hope class, which made up 47.9% of
the data. These findings highlight the importance of applying targeted sampling strategies to improve
minority class representation and, ultimately, enhance overall model performance in detecting nuanced
expressions of hope.</p>
      <p>While our methodology shows impressive baseline performance, there remain several notable
limitations that must be discussed. For instance, the current prompting strategies fall short in capturing
subtle sarcasm and context-dependent expressions of hope. To overcome these issues, we propose three
avenues for future exploration. First, implementing chain-of-thought prompting might allow models
to break down complex utterances into distinct contextual signals prior to making a final decision.
Second, employing an ensemble of transformer models—by combining predictions from architectures
such as DeBERTa, BETO, and Llama—could leverage their unique strengths, especially in low-resource
scenarios for sarcasm detection. Finally, integrating culture-aware data augmentation techniques may
help reduce dialectical biases in Spanish hope speech by utilizing region-specific lexicons to boost
generalization. These limitations underscore the need for further research into improved fine-tuning
methods and innovative data augmentation practices to elevate overall performance.</p>
      <sec id="sec-5-1">
        <title>5.1. Conclusion</title>
        <p>In summary, the system showed robust performance in detecting hope speech on English and Spanish
social media posts, with binary F1 scores around 0.83–0.84 achieved by transformer-based models.
However, accurately distinguishing the subtle diferences among hope expressions remains a challenge,
especially when implicit signals mix with sarcasm. In our experiments, the LLaMA model was also
employed for emotion detection. While it demonstrated promising potential, its performance has not
yet surpassed that of the established transformer architectures. This indicates that despite LLaMA’s
advanced capabilities, further refinement in instruction tuning and architectural adjustments is needed
for it to efectively capture subtle linguistic signals in hope speech detection. Future work should aim
to train models that understand context, adjust to diferent dialects, and combine the strengths of both
transformer and LLaMA methods to boost accuracy.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgements</title>
      <p>This research was supported by the VNUHCM-University of Information Technology’s Scientific
Research Support Fund.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>Generative AI tools were employed only to enhance linguistic clarity. All data, analyses, and conclusions
were conceived, executed, and validated by the authors. The authors retain full accountability for every
aspect of the manuscript’s content.
[9] J. Cañete, G. Chaperon, R. Fuentes, J.-H. Ho, H. Kang, J. Pérez, Spanish pre-trained bert model and
evaluation data, arXiv preprint arXiv:2308.02976 (2023).
[10] A. Gutiérrez-Fandiño, J. Armengol-Estapé, M. Pàmies, J. Llop-Palao, J. Silveira-Ocampo, C. P.</p>
      <p>Carrino, A. Gonzalez-Agirre, C. Armentano-Oller, C. Rodriguez-Penagos, M. Villegas, Spanish
language models, arXiv preprint arXiv:2107.07253 (2021).
[11] T. Brown, B. Mann, N. Ryder, M. Subbiah, J. D. Kaplan, P. Dhariwal, A. Neelakantan, P. Shyam,
G. Sastry, A. Askell, et al., Language models are few-shot learners, Advances in neural information
processing systems 33 (2020) 1877–1901.
[12] H. Touvron, T. Lavril, G. Izacard, X. Martinet, M.-A. Lachaux, T. Lacroix, B. Rozière, N. Goyal,
E. Hambro, F. Azhar, et al., Llama: Open and eficient foundation language models, arXiv preprint
arXiv:2302.13971 (2023).
[13] A. Ngo, H. T. H. Tran, Zootopi at hope2023@ iberlef: Is zero-shot chatgpt the future of hope
speech detection?, in: IberLEF@ SEPLN, 2023.
[14] E. J. Hu, Y. Shen, P. Wallis, Z. Allen-Zhu, Y. Li, S. Wang, L. Wang, W. Chen, et al., Lora: Low-rank
adaptation of large language models., ICLR 1 (2022) 3.
[15] S. Butt, F. Balouchzahi, M. Amjad, S. M. Jiménez-Zafra, H. G. Ceballos, G. Sidorov, Overview
of polyhope at IberLEF 2025: Optimism, expectation or sarcasm?, Procesamiento del Lenguaje
Natural (2025).
[16] S. Butt, F. Balouchzahi, A. I. Amjad, M. Amjad, H. G. Ceballos, S. M. Jiménez-Zafra, Optimism,
expectation, or sarcasm? multi-class hope speech detection in spanish and english, 2025.
[17] G. Sidorov, F. Balouchzahi, L. Ramos, H. Gómez-Adorno, A. Gelbukh, MIND-HOPE: Multilingual</p>
      <p>Identification of Nuanced Dimensions of HOPE, Technical Report, 2024.
[18] F. Balouchzahi, S. Butt, M. Amjad, G. Sidorov, A. Gelbukh, UrduHope: Analysis of hope and
hopelessness in urdu texts, Knowledge-Based Systems 308 (2025) 112746.
[19] G. Sidorov, F. Balouchzahi, S. Butt, A. Gelbukh, Regret and hope on transformers: An analysis of
transformers on regret and hope speech detection datasets, Applied Sciences 13 (2023) 3983.
[20] D. García-Baena, F. Balouchzahi, S. Butt, M. García-Cumbreras, A. L. Tonja, J. A. García-Díaz,
S. M. Jiménez-Zafra, Overview of hope at IberLEF 2024: Approaching hope speech detection in
social media from two perspectives, Procesamiento del Lenguaje Natural 73 (2024) 407–419.
[21] D. García-Baena, M. García-Cumbreras, S. M. Jiménez-Zafra, J. A. García-Díaz, V. G. Rafael, Hope
speech detection in spanish: The LGTB case, Language Resources and Evaluation (2023) 1–31.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>González-Barba</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Chiruzzo</surname>
            ,
            <given-names>S. M.</given-names>
          </string-name>
          <string-name>
            <surname>Jiménez-Zafra</surname>
          </string-name>
          ,
          <article-title>Overview of IberLEF 2025: Natural Language Processing Challenges for Spanish and other Iberian Languages, in: Proceedings of the Iberian Languages Evaluation Forum (IberLEF 2025), co-located with the 41st Conference of the Spanish Society for Natural Language Processing (SEPLN 2025), CEUR-WS</article-title>
          . org,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , Bert:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          ,
          <source>in: Proceedings of the NAACL</source>
          ,
          <year>2019</year>
          , pp.
          <fpage>4171</fpage>
          -
          <lpage>4186</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Conneau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Khandelwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Chaudhary</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Wenzek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Guzmán</surname>
          </string-name>
          , E. Grave,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          ,
          <article-title>Unsupervised cross-lingual representation learning at scale</article-title>
          , arXiv preprint arXiv:
          <year>1911</year>
          .
          <volume>02116</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>T.</given-names>
            <surname>Pires</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Schlinger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Garrette</surname>
          </string-name>
          ,
          <article-title>How multilingual is multilingual bert?</article-title>
          , arXiv preprint arXiv:
          <year>1906</year>
          .
          <volume>01502</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J. L. D.</given-names>
            <surname>Olmedo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Vázquez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. P.</given-names>
            <surname>Álvarez</surname>
          </string-name>
          ,
          <article-title>I2c-huelva at hope2023@ iberlef: Simple use of transformers for automatic hope speech detection (</article-title>
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>Rodríguez-García</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Riaño-Martínez</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Montalvo-Herranz</surname>
          </string-name>
          ,
          <article-title>Urjc-team at hope2023@ iberlef: Multilingual hope speech detection using transformers architecture</article-title>
          ., in: IberLEF@ SEPLN,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S. J.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Paas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. S.</given-names>
            <surname>Ahn</surname>
          </string-name>
          ,
          <article-title>Transformer transfer learning emotion detection model: synchronizing socially agreed and self-reported emotions in big data</article-title>
          ,
          <source>Neural Computing and Applications</source>
          <volume>35</volume>
          (
          <year>2023</year>
          )
          <fpage>10945</fpage>
          -
          <lpage>10956</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>P.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gao</surname>
          </string-name>
          , W. Chen, Deberta:
          <article-title>Decoding-enhanced bert with disentangled attention</article-title>
          , arXiv preprint arXiv:
          <year>2006</year>
          .
          <volume>03654</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>