<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Dandys-de-BERTganim at EXIST 2025: a Multi-task Learning Architecture for Sexism Identification</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Marc Hurtado</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Aleixandre Tarrasó</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Universitat Politècnica de València</institution>
          ,
          <addr-line>Camí de Vera, sn 46022 València</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>This paper presents Dandys, our system developed for the EXIST 2025 shared task on sexism identification in social media at CLEF 2025. The challenge comprises three interconnected tasks: binary sexism detection, source intention classification, and sexism type categorization. To address these tasks, we adopt a multi-task learning architecture with language-specific transformers for English and Spanish tweets, integrating demographic information from annotators as contextual signals. We enhance model generalization through data augmentation techniques such as back-translation and a punctuation-based augmentation method. Furthermore, we introduce a soft-labeling data reader to better reflect annotation disagreement, aligning with the Learning with Disagreement paradigm. Our results demonstrate the efectiveness of leveraging task interdependence, soft supervision, and multilingual modeling for addressing complex sociolinguistic classification problems. Given good performance across all tasks, our architecture is validated as both robust and efective.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Sexism Detection</kwd>
        <kwd>Social Media</kwd>
        <kwd>Multi-task Learning</kwd>
        <kwd>Data Augmentation</kwd>
        <kwd>Annotator Demographics</kwd>
        <kwd>Soft Labeling</kwd>
        <kwd>Multilingual NLP</kwd>
        <kwd>Learning with Disagreement</kwd>
        <kwd>EXIST 2025</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
    </sec>
    <sec id="sec-2">
      <title>2. Dandys-de-BERTganim System</title>
      <p>In this section, we provide a comprehensive description of the components, design choices, and
implementation details of the Dandys system that we developed for the EXIST 2025 shared task.</p>
      <sec id="sec-2-1">
        <title>2.1. Data Augmentation</title>
        <p>Data augmentation is crucial for improving model robustness and generalization, especially in
low-resource or highly imbalanced settings like sexism detection on social media, where linguistic
variability and noisy user-generated content can hinder performance. To increase the volume
and diversity of our training data, we applied two complementary data augmentation strategies:
back-translation and punctuation-based augmentation.</p>
        <p>
          Back-translation is a technique that involves translating a piece of text from its original language into
another language and then translating it back into the original language. This process helps generate
new linguistic variants while preserving the semantic content of the text. For both English and Spanish,
we used the state-of-the-art Helsinki-NLP opus-mt models[
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] to translate the tweets. Figure 1 shows
a schema of the Back-translation strategy and Table 1 presents an example of the back-translation
process. Moreover, due to tweets being written in Spanish and English, the first translation is also used
to augment the dataset of the other language.
        </p>
        <p>
          In addition, we implemented a punctuation-based augmentation method known as AEDA[
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] (An
Easier Data Augmentation). This technique involves inserting random punctuation marks into the
sentence without altering its meaning. Figure 2 shows a schema of the combination of Back-translation
and AEDA-Augmented strategies. We used a punctuation insertion ratio of approximately 0.3, meaning
that punctuation was added before about 30% of the words. This approach helped enrich the syntactic
variability of the dataset, which in turn supports better generalization in model training. Table 2
presents an example of the AEDA-Augmented process.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Dual dataset processing</title>
        <p>During our experiments, we found it necessary to implement two distinct data processing components
to address both the hard (single-label) and soft (distributional-label) subtasks:
• Hard-Labeling Data Reader Designed for the hard subtask, this reader applies a majority-vote
strategy: each tweet is assigned exactly one label corresponding to the class chosen by the
majority of annotators.</p>
        <p>– Advantages: Simplicity and speed in label generation.
– Limitations: Discards annotation disagreement, potentially introducing label noise for
ambiguous or context-dependent content.
• Soft-Labeling Data Reader Developed for the soft subtask, this reader preserves the full
distribution of annotator responses. Instead of collapsing all annotations into a single class, it
returns a probability vector reflecting class agreement. For example:</p>
        <p>︀[ 0.40Direct, 0.30Reported, 0.30Judgmental︀] .
– Advantages: Captures the subjectivity and uncertainty inherent in social media discourse.</p>
        <p>Also, enables the model to learn from disagreement patterns, improving probability
calibration and robustness.</p>
        <p>Both readers share the same preprocessing pipeline (normalization, tokenization, emoji handling),
but difer in label construction:
• When the hard reader is used, the model is trained using standard cross-entropy loss on one-hot
label vectors, which include an additional “ambiguous” class to represent uncertain cases.
• When the soft reader is used, the model is trained using Kullback–Leibler divergence loss
(KL</p>
        <p>DivLoss) on distributional label vectors, following the Learning with Disagreement paradigm.</p>
        <p>This dual-reader design allows our multi-task architecture to be trained in two complementary
modes:
1. With hard labels (including the extra ambiguous class) for maximizing precision on clear-cut
examples.
2. With soft labels and KL divergence supervision for capturing nuances and improving robustness
on ambiguous cases.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Annotators Information</title>
      <p>
        Another important component of our system involves leveraging the demographic metadata of the
annotators[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Each tweet in the dataset is labeled by six diferent individuals, whose interpretations
are inevitably shaped by their personal background[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. To better capture this annotator-specific
context, we designed a module that encodes demographic information—specifically, gender,
age, ethnicity, education level, and nationality—into fixed-size vector representations. These
embeddings aim to model the diversity of perspectives that arise from diferent social and cultural positions.
      </p>
      <p>To achieve this, we precomputed a set of 78,900 sentence-level embeddings, each representing a
unique combination of annotator demographic attributes, as shown in Table 3. Each sentence was then
encoded into a dense vector using the LaBSE sentence transformer model and stored in a lookup table
for eficient retrieval during training and inference.</p>
      <p>During training and inference, for each tweet, we retrieve the demographic descriptions of its six
annotators and extract the corresponding embeddings from the table. These six vectors are then
averaged to obtain a single representation that summarizes the annotator context for that tweet. This
aggregated embedding is used as an additional input to the classifier, providing rich contextual signals
that help the model to interpret annotation patterns more accurately—particularly in subjective or
ambiguous cases. This process is shown in Figure 3</p>
      <p>However, in this edition of the task, we observed that incorporating annotator embeddings did not
lead to significant performance improvements. Unlike previous years, where annotator demographics
were sometimes unevenly distributed—resulting in patterns or biases that the model could exploit—this
year’s dataset presents a more balanced and diverse allocation of annotators across tweets. As a result,
the demographic information of annotators does not ofer strong predictive cues about their labeling
tendencies. This limits the added value of explicitly modeling their profiles, although we believe such
contextual representations remain a promising direction for future work, especially in scenarios where
annotator bias is more prominent or unevenly distributed.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Model Architecture and Fine-tuning Process.</title>
      <p>In this section, we explain the architecture of the model. First, we will focus on the core architecture,
and the diferent steps that are carried out within it. Later, we will explain the multitasking architecture
implemented and how we have applied it to the diferent subtasks.</p>
      <p>The backbone encoder shown in Figure 4 is a stack of Transformer layers that processes the input
text to generate contextual embeddings. It begins with an input embedding layer that combines token
embeddings with positional information. This is followed by multiple identical layers, each containing
a multi-head self-attention mechanism and a feed-forward neural network. Residual connections
and layer normalization are applied after each sub-layer to stabilize and enhance learning. As the
input passes through these layers, the model captures increasingly complex patterns and relationships
between words, resulting in a rich contextual representation of the input sequence.</p>
      <p>
        Our model architecture is based on a multi-task[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ][
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] design that simultaneously handles all three
classification subtasks. We developed two separate but identical model instances: one for English
tweets and another for Spanish tweets. The English model is based on the xlm-roberta-base-sentiment
transformer [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], while the Spanish model uses robertuito-sentiment-analysis [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
      </p>
      <p>In addition, for subtask 2 (source intention classification), we used a single multilingual model
twitter-xlm-roberta-large-2022 for both Spanish and English. This choice was motivated by the
benefits of training on a combined dataset, which provided more annotated examples per class
and improved generalization. Two of our submissions relied exclusively on this multilingual
model, trained on a broader set of tweets spanning both languages. This approach allowed us to
leverage a larger and more diverse training set, resulting in slight performance gains, particularly on
underrepresented categories and borderline cases. The multilingual setup proved especially efective
when Spanish and English data were merged, ofering richer supervision and more robust generalization.</p>
      <p>Once a tweet is processed by its respective transformer, the resulting tweet the embedding of the
[CLS] token, is concatenated with the average annotator embedding. This joint representation is then
passed through a shared encoder, which feeds into three separate output heads—each responsible for one
of the classification tasks. Each output head consists of a dense layer with ReLU activation followed by
another dense layer with a softmax or sigmoid activation function to produce the final class probabilities.</p>
      <p>This architecture ofers several advantages: by using a shared encoder, we reduce the total
number of parameters, making the model more eficient. The multi-task learning setup promotes a
shared representation learned across tasks, which can lead to improved performance by leveraging
complementary information. Finally, the dual-model and multilingual strategies allow for both
language-specific fine-tuning and cross-lingual knowledge transfer. A schema of the model architecture
is shown in Figure 5.</p>
      <p>
        A critical component in obtaining strong performance is the selection of key hyperparameters, such
as learning rate, batch size, weight decay, dropout rate, and class imbalance handling strategies. To this
end, we employed Optuna[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], an automatic hyperparameter optimization framework. Optuna enabled
eficient exploration of the search space through Bayesian optimization and pruning, allowing us to
identify optimal configurations for each model variant with reduced computational cost.
      </p>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
      <p>The following tables present the oficial competition results for the EXIST2025 shared task on sexism
identification. We compare our Dandys system against the gold-standards and the subtask winner
across all three subtasks in both the soft (distributional labels) and hard (single-label) settings. We
report Information Content Metrics (ICM), normalized ICM (ICM Norm), and task-specific performance
measures: cross-entropy loss for soft subtasks, F1-score for binary detection (Task 1), and macro-F1 for
the multiclass subtasks (Tasks 2 and 3). Notably, our performance on Task 2 (Source Intention) is good,
demonstrating that our multi-task architecture and augmentation strategies are particularly efective
for modeling intent.</p>
      <p>Team-Ranking ICM-Hard ICM-Hard Norm</p>
      <p>F1 YES</p>
      <sec id="sec-5-1">
        <title>5.1. Analysis of the results</title>
        <p>When comparing our system, Dandys-de-BERTganim, to the top two teams in the EXIST2025 challenge
GrootWatch_1 and Mario_1, we observed both strengths and areas for improvement.</p>
        <p>In soft settings, our model generally had lower ICM-Soft and normalized scores than GrootWatch_1.
For example, in Subtask 1, we scored 0.6575 vs. their 1.0600, and in Subtask 3, our ICM-Soft dropped to
–8.7671. However, our Cross Entropy was consistently better, indicating well-calibrated probability
outputs. In Subtask 1 Soft, we achieved 0.7964 vs. GrootWatch_1’s 0.8893, and in Subtask 2 Soft, we
outperformed them again (1.3820 vs. 1.7711) while ranking 2nd overall.</p>
        <p>In hard settings, our best result was in Subtask 2, where we ranked 4th, with a Macro F1 of 0.5522,
close to Mario_1’s 0.5692, the top performer. This suggests strong classification performance in source
intention. However, in Subtask 1 Hard, we lagged behind, with an F1 YES of 0.7548 vs. Mario_1’s 0.8167,
and in Subtask 3 Hard, our F1 of 0.5827 was lower than GrootWatch_1’s 0.6305.</p>
        <p>Overall, while the top two teams led in label agreement and F1 scores, our system showed competitive
results in probabilistic modeling and source intention classification. Future work should focus on
improving alignment with annotations and boosting accuracy in stricter evaluation settings.</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Conclusions and Future Work</title>
        <p>Across all subtasks, the Dandys system demonstrates competitive performance. While there remains
room for improvement in Task 1 and Task 3, our approach shows robustness in handling ambiguous
and context-dependent examples. Most notably, in Task 2 (Source Intention), Dandys achieves high
normalized ICM scores in both soft and hard settings, underscoring the benefits of our dual-reader
design, data augmentation strategies, and multi-task learning framework for modeling nuanced intent
in social media posts.</p>
        <p>As future work, we plan to explore alternative strategies for integrating annotator demography
information into the model, including position-agnostic embeddings and attention mechanisms for
demographic information.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>During the development of this research work, the authors employed multiple generative artificial
intelligence (AI) systems to enhance various aspects of the manuscript preparation process.
Specifically, large language models (LLMs) including ChatGPT (OpenAI), DeepSeek (DeepSeek AI), Claude
(Anthropic), and Gemini (Google DeepMind) were utilized for several critical purposes:
• Writing Enhancement: These tools assisted in improving grammatical accuracy, stylistic
coherence, and overall readability of the manuscript.
• Comprehensive Literature Review: AI systems were used to conduct extensive searches of
relevant literature, identify key research gaps, and summarize complex academic sources.
• Code Optimization: Where applicable, AI-assisted suggestions were implemented to refine
computational methods, debug algorithms, and improve eficiency in software implementations.</p>
      <p>The authors rigorously reviewed, fact-checked, and edited all AI-generated content to ensure accuracy,
originality, and alignment with scholarly standards. Additionally, all critical insights, theoretical
contributions, and final conclusions remain the original work of the authors. The use of these tools was
strictly supplementary, and full responsibility for the research content, ethical considerations, and final
presentation lies with the authors.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>L.</given-names>
            <surname>Plaza</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. C. de Albornoz</surname>
            , I. Arcos,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Spina</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Amigó</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Gonzalo</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Morante</surname>
          </string-name>
          , Overview of EXIST 2025:
          <article-title>Learning with Disagreement for Sexism Identification and Characterization in Tweets, Memes, and TikTok Videos (Extended Overview)</article-title>
          , in: G. Faggioli,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ferro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          , D. Spina (Eds.),
          <source>CLEF 2025 Working Notes</source>
          ,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>L.</given-names>
            <surname>Plaza</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. C. de Albornoz</surname>
            , I. Arcos,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Spina</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Amigó</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Gonzalo</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Morante</surname>
          </string-name>
          , Overview of EXIST 2025:
          <article-title>Learning with Disagreement for Sexism Identification and Characterization in Tweets, Memes, and TikTok Videos</article-title>
          , in: J.
          <string-name>
            <surname>C. de Albornoz</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Gonzalo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Plaza</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. G. S. de Herrera</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Mothe</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Piroi</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Spina</surname>
          </string-name>
          , G. Faggioli, N. Ferro (Eds.),
          <source>Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Sixteenth International Conference of the CLEF Association (CLEF</source>
          <year>2025</year>
          ),
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>J.</given-names>
            <surname>Tiedemann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Aulamo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Bakshandaeva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Boggia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.-A.</given-names>
            <surname>Grönroos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Nieminen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Raganato</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Scherrer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Vázquez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Virpioja</surname>
          </string-name>
          ,
          <article-title>Democratizing neural machine translation with opus-mt, arXiv preprint (</article-title>
          <year>2022</year>
          ). URL: https://arxiv.org/pdf/2212.
          <year>01936</year>
          . arXiv:
          <fpage>2212</fpage>
          .
          <year>01936</year>
          , preprint.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>A.</given-names>
            <surname>Karimi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Rossi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Prati</surname>
          </string-name>
          ,
          <string-name>
            <surname>AEDA:</surname>
          </string-name>
          <article-title>An easier data augmentation technique for text classification, arXiv preprint (</article-title>
          <year>2021</year>
          ). URL: https://arxiv.org/pdf/2108.13230. arXiv:
          <volume>2108</volume>
          .13230, preprint.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Y.-Z.</given-names>
            <surname>Fang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.-H.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-D.</given-names>
            <surname>Huang</surname>
          </string-name>
          , NYCU-NLP at EXIST 2024:
          <article-title>Leveraging Transformers with Diverse Annotations for Sexism Identification in Social Networks, in: Notebook for the EXIST Lab at CLEF</article-title>
          <year>2024</year>
          ,
          <year>2024</year>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3740</volume>
          /paper-93.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Z. L. N. Z. J. G. L. Shengnan</given-names>
            <surname>An</surname>
          </string-name>
          , Zexiong Ma,
          <article-title>Make your llm fully utilize the context</article-title>
          , arXiv preprint (
          <year>2024</year>
          ). URL: https://arxiv.org/pdf/2404.16811. arXiv:
          <volume>2404</volume>
          .
          <fpage>16811</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Yan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Mo</surname>
          </string-name>
          <article-title>, Multi-task learning for sentiment analysis with hard-sharing and task recognition mechanisms</article-title>
          ,
          <source>Information</source>
          <volume>12</volume>
          (
          <year>2021</year>
          )
          <article-title>207</article-title>
          . URL: https://www.mdpi.com/2078-2489/12/5/ 207. doi:
          <volume>10</volume>
          .3390/info12050207.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>O.</given-names>
            <surname>Sener</surname>
          </string-name>
          ,
          <string-name>
            <surname>V.</surname>
          </string-name>
          <article-title>Koltun, Multi-task learning as multi-objective optimization</article-title>
          ,
          <source>in: Advances in Neural Information Processing Systems</source>
          , volume
          <volume>31</volume>
          ,
          <year>2018</year>
          . URL: https://proceedings.neurips.cc/paper_files/ paper/2018/file/432aca3a1e345e339f35a30c8f65edce-Paper.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J.</given-names>
            <surname>Camacho-collados</surname>
          </string-name>
          , K. Rezaee,
          <string-name>
            <given-names>T.</given-names>
            <surname>Riahi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ushio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Loureiro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Antypas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Boisson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. Espinosa</given-names>
            <surname>Anke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. Martínez</given-names>
            <surname>Cámara</surname>
          </string-name>
          , et al.,
          <article-title>Tweetnlp: Cutting-edge natural language processing for social media</article-title>
          ,
          <source>Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing: System Demonstrations</source>
          (
          <year>2022</year>
          )
          <fpage>38</fpage>
          -
          <lpage>49</lpage>
          . URL: https://arxiv.org/abs/2206.14774.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>J. M. Pérez</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Rajngewerc</surname>
            ,
            <given-names>J. C.</given-names>
          </string-name>
          <string-name>
            <surname>Giudici</surname>
            ,
            <given-names>D. A.</given-names>
          </string-name>
          <string-name>
            <surname>Furman</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Luque</surname>
            ,
            <given-names>L. A.</given-names>
          </string-name>
          <string-name>
            <surname>Alemany</surname>
            ,
            <given-names>M. V.</given-names>
          </string-name>
          <string-name>
            <surname>Martínez</surname>
          </string-name>
          ,
          <article-title>pysentimiento: a python toolkit for opinion mining and social nlp tasks</article-title>
          ,
          <source>arXiv preprint arXiv:2106.09462</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>J. M. Pérez</surname>
            ,
            <given-names>D. A.</given-names>
          </string-name>
          <string-name>
            <surname>Furman</surname>
            ,
            <given-names>L. Alonso</given-names>
          </string-name>
          <string-name>
            <surname>Alemany</surname>
            ,
            <given-names>F. M.</given-names>
          </string-name>
          <string-name>
            <surname>Luque</surname>
          </string-name>
          ,
          <article-title>RoBERTuito: a pre-trained language model for social media text in Spanish, in: Proceedings of the Thirteenth Language Resources</article-title>
          and Evaluation Conference, European Language Resources Association, Marseille, France,
          <year>2022</year>
          , pp.
          <fpage>7235</fpage>
          -
          <lpage>7243</lpage>
          . URL: https://aclanthology.org/
          <year>2022</year>
          .lrec-
          <volume>1</volume>
          .
          <fpage>785</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M.</given-names>
            <surname>García-Vega</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Díaz-Galiano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>García-Cumbreras</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Del Arco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Montejo-Ráez</surname>
          </string-name>
          , S. JiménezZafra, E. Martínez Cámara,
          <string-name>
            <given-names>C.</given-names>
            <surname>Aguilar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Cabezudo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Chiruzzo</surname>
          </string-name>
          , et al.,
          <source>Overview of tass</source>
          <year>2020</year>
          :
          <article-title>Introducing emotion detection</article-title>
          ,
          <source>in: Proceedings of the Iberian Languages Evaluation Forum (IberLEF</source>
          <year>2020</year>
          )
          <article-title>Co-Located with 36th Conference of the Spanish Society for Natural Language Processing (SEPLN</article-title>
          <year>2020</year>
          ), Málaga, Spain,
          <year>2020</year>
          , pp.
          <fpage>163</fpage>
          -
          <lpage>170</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Optuna</given-names>
            <surname>Development</surname>
          </string-name>
          <string-name>
            <surname>Team</surname>
          </string-name>
          ,
          <article-title>Optuna: A hyperparameter optimization framework</article-title>
          ,
          <source>Documentation</source>
          ,
          <year>2025</year>
          . URL: https://optuna.readthedocs.io/en/stable/index.html, accessed: June 1,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>