<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>FHSTP@EXIST 2025 Benchmark: Sexism Detection with Transparent Speech Concept Bottleneck Models⋆</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Roberto Labadie-Tamayo</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Adrian Jaques Böck</string-name>
          <email>Adrian.Boeck@fhstp.ac.at</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Djordje Slijepčević</string-name>
          <email>Djordje.Slijepcevic@fhstp.ac.at</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Xihui Chen</string-name>
          <email>Xihui.Chen@fhstp.ac.at</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Andreas Babic</string-name>
          <email>Andreas.Babic@fhstp.ac.at</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Matthias Zeppelzauer</string-name>
          <email>Matthias.Zeppelzauer@fhstp.ac.at</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>@WillyTolerdoo @MeerRocio Yo es que soy “masculinista”..</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>St. Pölten University of Applied Sciences</institution>
          ,
          <addr-line>3100 St. Pölten</addr-line>
          ,
          <country country="AT">Austria</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <abstract>
        <p>Sexism has become widespread on social media and in online conversation. To help address this issue, the iffth Sexism Identification in Social Networks (EXIST) challenge is initiated at CLEF 2025. Among this year's international benchmarks, we concentrate on solving the first task aiming to identify and classify sexism in social media textual posts. In this paper, we describe our solutions and report the results for all the three subtasks: Subtask 1.1 - Sexism Identification in Tweets , Subtask 1.2 - Source Intention in Tweets, and Subtask 1.3 - Sexism Categorization in Tweets. We implement three models to address each subtask which constitute three individual runs, i.e., Speech Concept Bottleneck Model (SCBM), Speech Concept Bottleneck Model with Transformer (SCBMT) and a fine-tuned XLM-RoBERTa transformer model. SCBM uses descriptive adjectives as human-interpretable bottleneck concepts. SCBM leverages large language models (LLMs) to encode input texts to a human-interpretable representation of descriptive adjectives, which is then used to train a light-weight classifier for downstream tasks. SCBMT extends SCBM by fusing the adjective-based representation with the contextual embeddings computed by transformers to balance interpretability and classification performance. Beyond the competitive results, these two models ofer fine-grained explanations at both the instance level (local) and class level (global). We also investigate how additional metadata available in the benchmarks, e.g., annotators' demographic profiles, can be leveraged in our models. For the binary classification task (Subtask 1.1), XLM-RoBERTa, fine-tuned on the provided data augmented with datasets from previous years, ranks 6th for English and Spanish content and 4th for English content in the Soft-Soft evaluation. Our SCBMT model achieves 7th place for English and Spanish content and 6th for Spanish content in the Soft-Soft evaluation.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Sexism Detection</kwd>
        <kwd>Sexism Identification</kwd>
        <kwd>Social Media Retrieval</kwd>
        <kwd>Concept Bottleneck Models</kwd>
        <kwd>Transformer Models</kwd>
        <kwd>Natural Language Processing</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>We employ three distinct machine learning (ML) models to address each subtask and provide three
diferent experimental runs for each. Our first approach utilizes the Speech Concept Bottleneck Model
(SCBM), a human-interpretable model that leverages large language models (LLMs) to map input text into
an adjective-based representation. Originally developed for recognizing online hate and counter speech,
SCBM ofers both global and local explainability. Each dimension of the intermediate representation
reflects the degree to which a specific adjective captures an emotion or subjective opinion expressed in
the input text. A lightweight classifier is then trained on these representations for downstream prediction
tasks. Compared to concept representations based on topics or linguistic patterns, using adjectives
as bottleneck concepts significantly improves the interpretability of the model’s outputs. Our second
model, the Speech Concept Bottleneck Model with Transformer (SCBMT), extends SCBM by combining
the adjective-based representations with transformer-generated embeddings. This fusion enables the
model to capture a more comprehensive range of textual features by integrating interpretable and
contextual information. Finally, we fine-tune a multilingual RoBERTa transformer model as a baseline
to benchmark the performance of our proposed methods.</p>
      <p>The structure of the paper. We describe in detail our methodology used to address the challenge,
including the developed models and evaluation settings (e.g., datasets, concept list), in Section 2. The
results are presented in Section 3.1, followed by the discussion and conclusions in Section 4.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Methodology and Evaluation</title>
      <p>With the three models employed, we conducted three independent runs for each subtask of the tweet
classification challenge of the benchmark. In this section, we describe the datasets used, the developed
ML models, and the approaches applied for each submitted run of the subtasks.</p>
      <sec id="sec-2-1">
        <title>2.1. Datasets</title>
        <p>In addition to the dataset (EXIST2025) of this year’s benchmark, we utilized data provided in previous
years’ benchmarks (EXIST2021-EXIST2024). The combination increases the scale of training data, and
thereby improves model robustness and generalizability.</p>
        <p>
          EXIST2025 Dataset. The EXIST2025 dataset [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] contains over 10,000 text posts from X/Twitter, written
in both English and Spanish, with a balanced distribution between the two languages. The dataset is split
into three subsets: training (6,920 posts), development (1,038 posts), and testing (2,076 posts). Each post is
annotated by six individuals, and demographic information about the annotators (e.g., gender, age group,
ethnicity, education level, and country) is also provided. Each post is assigned three labels corresponding
to the dataset’s three subtasks: Sexism Identification , Source Intention, and Sexism Categorization. Subtask
1.1 – Sexism Identification is a binary classification task in which each post is labeled as either SEXIST or
NON-SEXIST. Subtask 1.2 – Source Intention is a multiclass classification task that aims to determine the
intention behind sexist content, with possible labels being DIRECT, REPORTED, or JUDGEMENTAL.
Subtask 1.3 – Sexism Categorization is a multi-label classification task that identifies specific types
of sexist content. The possible categories include: IDEOLOGICAL-INEQUALITY,
STEREOTYPINGDOMINANCE, OBJECTIFICATION, SEXUAL-VIOLENCE, and MISOGYNY-NON-SEXUAL-VIOLENCE.
EXIST2024 Dataset. The dataset from the EXIST2024 [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] benchmark includes posts from X/Twitter
and Gab. We used thetraining set, which consists of 6,920 instances in total, i.e., 3,260 in English and
3,660 in Spanish, and the development set, which includes 1,038 samples. The test set was not used due
to missing annotations. The class definitions for the tasks of Sexism Identification, Source Intention,
and Sexism Categorization are the same as those in the EXIST2025 dataset.
        </p>
        <p>EXIST2022 Dataset.</p>
        <p>
          We used the training dataset from the EXIST2022 benchmark [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]. This dataset includes posts from
X/Twitter and Gab, with a total of 11,345 instances: 5,644 in English and 5,701 in Spanish. The class
definitions for the tasks of Sexism Identification and Sexism Categorization are the same as those in the
EXIST2025 dataset. The Source Intention task was not included in the benchmark of the EXIST2022
datasets.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Lexicon of Descriptive Adjectives for Concept Bottleneck Models</title>
        <p>
          Considering the specific characteristics of sexist content, we propose a lexicon of descriptive adjectives
as the set of concepts for our SCBM and SCBMT models. We automatically generate the lexicon by
prompting the LLM GPT-o3-mini [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. For each subtask, we provide the LLM with the task definition
and example instances, as outlined in [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ], and ask it to generate 50 adjectives relevant for distinguishing
between the classes of the respective classification task. Thus, the prompts are constructed with
incontext learning, consisting of two parts. The first part gives a short description of the classification
task, while the second part lists the definitions and examples of the class labels. In the following, we
demonstrate the prompt used for Task 1.1 as an example while the other two prompts can be found in
Appendix A:
Provide me with 50 adjectives that can be used to describe and distinguish classes in a binary classification
task where systems must decide whether or not a given tweet is sexist. The following classes are defined:
• Sexist, as in: “Woman driving, be careful!”
• Non-sexist, as in: “Just saw a woman wearing a mask outside spank her very tightly leashed dog and
        </p>
        <p>I gotta say I love learning absolutely everything about a stranger in a single instant.”
After combining the adjectives from all three subtasks and removing duplicates, we obtained a final
set of 132 adjectives (see Table 1).
2.3. Models
aggressive
assaultive
body-focused
commodifying
crude
denouncing
devaluing
dismissive
eroticizing
harassing
inegalitarian
invidious
misandric
objectifying
paternalistic
prejudiced
recounting
repugnant
scurrilous
snide
traditionalist
victimizing
androcentric
beauty-obsessed
categorical
condemning
cruel
depersonalizing
diminishing
disparaging
essentialist
hateful
inferiorizing
judgmental
misogynistic
observational
patriarchal
provocative
reductive
repulsive
sexist
stereotypical
undermining
vilifying
antagonistic
belittling
censuring
condescending
degrading
derisive
disapproving
disrespectful
exclusionary
homogenizing
inflammatory
lecherous
mocking
ofensive
patronizing
rapacious
regressive
reviling
sexualizing
stereotyping
unjust
violence-inciting
antifeminist
belligerent
chauvinistic
contemptuous
dehumanizing
derogatory
discriminatory
documenting
exploitative
hostile
insulting
lustful
moralizing
oppressive
pejorative
rape-advocating
remonstrative
rude
sexually charged
superficial
venomous
vitriolic
anti-egalitarian
biased
chronicling
crass
demeaning
descriptive
disdainful
documenting
factual
ignorant
intolerant
marginalizing
narrow-minded
overgeneralizing
persecutory
reactionary
reporting
scathing
shallow
superiority-minded
verifiable
vituperative
We implement three models to solve the tasks. The first two models are based on the method we
proposed for hate and counter speech recognition. With its promising performance, we apply two
variants of the approach to this new domain of sexism detection. In addition, we employ the multilingual
transformer XLM-RoBERTa as a baseline.</p>
        <sec id="sec-2-2-1">
          <title>2.3.1. Speech Concept Bottleneck Model (SCBM)</title>
          <p>SCBM is a deep learning model which we proposed based on the approach of concept bottleneck
models (CBM) for text classification. Unlike prior models relying on transformers, it provides
humaninterpretable explanations for the predictions in terms of the most related descriptive adjectives. The
model is the first to use adjectives as bottleneck concepts to represent input texts, capturing the
underlying emotions, intentions, and semantics. The SCBM model consists of two sequential steps
implemented by two modules: concept evaluation and text classification .</p>
          <p>
            Concept evaluation. This module is used to encode the input text into a representation according
to a list of adjectives. For each adjective in Table 1, we prompt an LLM to compute a relevance score
between 0 and 1, which evaluates how well the adjective describes the text. Specifically, we query
a frozen, 8-bit quantized, pre-trained instance of Llama3.1-8b-instruct [
            <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
            ] with a simple prompt:
“Tell me if the adjective [adjective] describes the content of the following text: [text]?”. Considering the
randomness of LLMs’ responses, we calculate the relevance score as the marginalized probability of
the LLM generating a response starting with a positive afirmative word. Specifically, we construct a
set of words (tokens) that may appear in the prefix of an afirmative response, including ‘Yes’, ’Si’, and
their variants such as ’_yes’, ‘YES’. For each word in the set, we obtain the probability for the employed
LLM to generate a response starting with that word. Then we add up all probabilities of the tokens
as the relevance score of the given adjective to the input text. The output of this module is a vector
representation, where each dimension corresponds to the relevance score of an adjective in the list.
Text classification. This module takes the representations of adjective concepts as input and predicts
the label(s) of the specific task. Instead of an one-layer perceptron used in previous CBM-based models,
we introduce a relevance gate to further dynamically adjust the importance of each adjective to the
ifnal decision. The adjusted representation is then given to a multi-layer perceptron (MLP).
          </p>
          <p>To train SCBM, we leverage the EXIST 2025 data and employ a fixed learning rate of 2e-3 with
RMSProp, optimized over 300 training epochs. Early stopping is applied based on macro-F1 performance
on the validation set.</p>
        </sec>
        <sec id="sec-2-2-2">
          <title>2.3.2. Speech Concept Bottleneck Model with Transformer (SCBMT)</title>
          <p>
            SCBMT is an extended variant of the SCBM model resulting from fusing the adjective-based
representation with an embedding of the input text obtained from a transformer model. The aim is to combine two
complementary sources of information, considering that transformer-based models are well known for
their capability to capture semantics in the text, while our concept bottleneck representation provides
more intention-oriented information about the text. In our implementation, we use the multilingual
transformer, XLM-RoBERTa-large [
            <xref ref-type="bibr" rid="ref11 ref12">11, 12</xref>
            ]. To align the adjective representation with the feature space
of the transformer embeddings, we apply a learned linear projection before concatenation. The SCBMT
is trained end-to-end to predict the label(s) of the text defined in the specific task. For training SCBMT,
we use a fixed learning rate of 1e-5 for all transformer blocks and MLP layers, with optimization
performed using RMSProp over 16 epochs.
          </p>
        </sec>
        <sec id="sec-2-2-3">
          <title>2.3.3. Fine-tuned XLM-RoBERTa</title>
          <p>We fine-tune XLM-RoBERTa-large on the EXIST2025 dataset. In the fine-tuning process, we additionally
incorporate data from previous EXIST benchmarks (EXIST2022 and EXIST2024) to increase the amount
of training data and thus aim to improve generalizability. This fine-tuned XLM-RoBERTa model serves
as a transformer-based baseline. For all three subtasks, we apply the same training parameters, with
slight diferences in the datasets used. The model is fine-tuned over three epochs with a learning rate of
3e-5, using a linear learning rate scheduler with 500 warm-up steps, weight decay of 0.01, and a batch
size of 8. Early stopping is applied based on macro-F1 performance on the validation set.</p>
        </sec>
      </sec>
      <sec id="sec-2-3">
        <title>2.4. Experiments</title>
        <p>In this section, we present the experimental settings which lead to the three submitted experimental
runs for each subtask of Task 1. Each run explores a distinct ML model. Specifically, we employ SCBM
and SCBMT in Run 1 and Run 2, respectively, while utilizing the fine-tuned XLM-RoBERTa model in
Run 3.</p>
        <p>Due to the specific features of the data in EXIST2025, we have two special considerations when
designing each experimental run. Recall that six annotators are hired to annotate each tweet, and
their demographic information is provided. Our first decision in implementing our experimental run is
whether and how to integrate this type of information. The second decision is related to the imbalanced
data distribution across classes. We evaluate the impact and select the best strategy to deal with this
imbalance.</p>
        <p>Demographic information. We explore whether incorporating demographic information can help
models better align with the diverse preferences and interpretations of annotators. To investigate
this, we evaluated the three ML models under two conditions: i) without incorporating demographic
information, and ii) with demographic information included. In the latter setting, we enriched the
prompts to the LLM with each annotator’s demographics as personas, i.e., gender, age, ethnicity,
educational background, and country. In our implementation, one example of this persona information
is ‘You are a man aged above 45 years old with latino ethnicity with a Bachelor’s degree coming from
Mexico’. Since each post is annotated by six individuals, we generated six distinct representations
per instance, treating them independently during training. At inference time, the model produced six
predictions, one for each annotator, and we used majority voting to determine the final prediction.
Mitigating class imbalance. Due to class imbalances presented in the dataset (primarily for the
multi-class and multi-label classification tasks), we further investigated the impact of undersampling
techniques as a mitigation strategy. These methods aim to reduce the dominance of majority classes by
selectively removing samples, thereby promoting a more balanced class distribution.</p>
        <p>In the following, we outline the specific model architecture used for each run, along with any
supplementary methods or configurations applied, such as demographic conditioning, data augmentation
(with EXIST datasets from previous years), or data undersampling, which contributed to the highest
performances on the development set. We notice that the performance of the deployed models varies
across diferent settings. Moreover, settings with more information incorporated do not always lead to
better results. According to the exploratory analysis with various settings, we only describe the settings
with the best performance in the following.</p>
        <sec id="sec-2-3-1">
          <title>2.4.1. Task 1.1: Sexism Identification in Tweets</title>
          <p>Run 1 (SCBM). We train the SCBM model without considering annotators’ demographic information.
Additionally, we do not apply undersampling, as the dataset for this subtask is balanced.
Run 2 (SCBMT). We train the SCBMT model with annotators’ demographic information incorporated
into the prompts as the persona provided to the LLM. For each instance, the model generates six
predictions (one per annotator). Majority voting is applied to determine the final output. Additionally,
no undersampling is need as the data are balanced.</p>
          <p>Run 3 (Fine-tuned XLM-RoBERTa). We fine-tune an XLM-RoBERTa model on both the training
data in EXIST2025 and those from previous years (EXIST2022 and EXIST2024). Annotator demographic
information does not improve performance, and no undersampling strategies are applied as the data
are balanced.</p>
        </sec>
        <sec id="sec-2-3-2">
          <title>2.4.2. Task 1.2: Source Intention in Tweets</title>
          <p>Run 1 (SCBM). We train the SCBM model without annotators’ demographic profiles during training.
To address class imbalance, we apply undersampling to the NON-SEXIST class, resulting in a more
balanced dataset.</p>
          <p>Run 2 (SCBMT). We train the SCBMT model and incorporate annotators’ demographic information as
the persona in the prompts provided to the LLM. Additionally, we do not apply undersampling.
Run 3 (Fine-tuned XLM-RoBERTa). We fine-tune an XLM-RoBERTa model on the EXIST2025 dataset
with undersampling the NON-SEXIST class. No annotators’ demographic information is integrated.</p>
        </sec>
        <sec id="sec-2-3-3">
          <title>2.4.3. Task 1.3: Sexism Categorization in Tweets</title>
          <p>Run 1 (SCBM). We train the SCBM model for multi-label classification without incorporating annotator
information or applying any undersampling strategies.</p>
          <p>Run 2 (SCBMT). We train the SCBMT model for multi-label classification without incorporating
annotator information or applying any undersampling strategies.</p>
          <p>Run 3 (Fine-tuned XLM-RoBERTa). We fine-tune an XLM-RoBERTa model on the EXIST2025 dataset
for multi-label classification, without applying undersampling or incorporating annotator information.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Results</title>
      <sec id="sec-3-1">
        <title>3.1. Classification Performances</title>
        <p>The results across all three subtasks show clear diferences in performance depending on the task and
language setting. In Subtask 1.1 (Table 2), XLM-RoBERTa performed best overall, particularly in the
Soft-Soft evaluation, achieving top ranks and the lowest Cross Entropy scores across languages. Our
best submission was in this subtask, where the XLM-RoBERTa model achieved the 4ℎ place in the
overall benchmark. In this subtask, SCBM and SCBMT performed almost the same in the Soft-Soft
evaluation, with SCBMT slightly outperforming SCBM. However, in the Hard-Hard evaluation, SCBM
performed significantly worse compared to SCBMT by about 2-6% in Cross Entropy and 4-10% in
ICM-Soft Norm across both languages.</p>
        <p>For Subtask 1.2 (Table 3), SCBMT showed stronger results compared to Subtask 1.1, where it achieved
the largest ICM-Soft scores in the Soft-Soft evaluation, ranking the 6ℎ place in the benchmark for
English. It also performed comparably well on the mixed English-Spanish dataset, securing the 7ℎ
place. XLM-RoBERTa was less consistent than in the previous task, particularly in the Hard-Hard
evaluation. The SCBM performed comparably to XLM-RoBERTa but overall considerably worse than
the combined approach SCBMT.</p>
        <p>Subtask 1.3 (Table 4) was the most challenging task across all models. SCBMT again yielded the
most reliable results, ranking 13th for Spanish in the Soft-Soft evaluation. SCBM performed better in
the Soft-Soft evaluation compared to XLM-RoBERTa but performed slightly worse in the Hard-Hard
evaluation.</p>
        <p>Overall, XLM-RoBERTa performed best in the binary classification setting of Subtask 1.1, especially
when fine-tuned with additional social media datasets from previous years. SCBMT performed better
for finer-grained categorization of Subtask 1.2 and 1.3 than the XLM-RoBERTa model (especially for
Spanish data). SCBM demonstrates the lowest performance in most cases, although its performance in
the Soft-Soft evaluation was comparable to that of XLM-RoBERTa (except for in Subtask 1.1).</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Interpretability of SCBM</title>
        <p>We observe that SCBM when applied alone is less efective than the other two models which benefit
from text representations calculated from encoder-based transformers. However, the use of adjective
concepts makes the model inherently interpretable. As previously mentioned, SCBM implements a
trainable relevance gate to further adjust the importance of adjective concepts and generate the final
input encoding for classification. This representation serves as the explanation of SCBM. In this section,
we will examine the quality of these adjective-based explanations from both the global and local level.</p>
        <p>To assess the local interpretability of SCBM, we manually examine the top 10 adjectives with the
largest relevance scores. In Table 5, we randomly select one correctly classified tweet from each
class defined in every subtask, and present the top 10 relevant adjectives. Across all three tasks, we
observe that in general, the SCBM produces semantically rich, domain-relevant adjective explanations.
The majority of these adjectives support the classification outcome. Furthermore, they also expose
underlying discursive features of sexist content.</p>
        <p>For instance, in Subtask 1.2, the model consistently diferentiates stylistic and intentional nuances.
For example, tweets labeled as DIRECT are associated with adjectives such as invasive, androcentric, and
violence-inciting, revealing overt and aggressive narrative strategies. In contrast, the JUDGEMENTAL
class leans on subtler descriptors like chauvinistic, provocative, and victim-blaming, pointing toward
implicit bias. Meanwhile, the REPORTED category is likely to be described by adjectives like critical,
incendiary, and moralizing, indicating second-hand commentary on sexist events, which aligns with the
meta-discursive nature of such content.</p>
        <p>To provide a global explanation for SCBM, we aggregate local explanations to identify the adjectives
that contribute the most to predicting specific classes. The global contribution of an adjective to a
particular class is computed as the mean value of its activations across all correctly classified instances in
the training set belonging to that class. Table 6 shows the global explanation obtained for the multi-label
and multi-class classification tasks. The global explanations provided by SCBM align strongly with the
patterns observed in local explanations, reinforcing the model’s consistency in how it maps adjective
concepts to classes. Frequently occurring adjectives like paternalistic, intolerant, and body-focused
appear both at the instance level and across entire classes, confirming that the model’s decisions are
grounded in stable, semantically relevant features.</p>
        <p>However, we also observe meaningful overlap between classes, particularly within ideologically
adjacent labels. For example, adjectives such as belligerent and invasive appear in both
SEXUALVIOLENCE and OBJECTIFICATION, while critical and androcentric are shared across
IDEOLOGICALINEQUALITY and STEREOTYPING-DOMINANCE. This reflects the fuzzy boundaries between sexism
subtypes and highlights the complexity of capturing such discourse through discrete labels.
@FlorenciaLagosN Andate a Cuba, yo te
pago los pasajes. . .</p>
        <p>Esta publicidad solo incluye a gais y
bisexuales hombres cisgénero...</p>
        <p>En nuestra provincia, existe una
importante brecha de género...</p>
        <p>Un hombre con plata es como una mujer
bonita...
@girldomf me gustaría follar a alguna de
ustedes...
@Maria_RuizG1 Ojalá no vote, porque las
mujeres no deberían votar :)...</p>
        <p>Explanation
victim-blaming, harassing, devaluing,
insulting, oppressive, sexist, misogynistic,
appearance-driven, predatory, hostile
androcentric, invasive, patriarchal,
violence-inciting, dehumanizing, abusive,
rape-advocating, assaultive, hostile, predatory
patriarchal, provocative, chauvinistic,
inflammatory, masculinist, simplistic,
homogenizing, repugnant, victim-blaming,
traditionalist
critical, incendiary, disempowering,
moralizing, inflammatory, biased,
denouncing, reductive, provocative,
depersonalizing
simplistic, paternalistic, essentialist,
stereotypical, traditionalist, narrow-minded,
intolerant, dehumanizing, exclusionary,
non-consensual
vitriolic, intolerant, devaluing, denigrating,
invasive, belittling, venomous, body-focused,
ofensive, inflammatory
superficial, body-focused, critical, devaluing,
rude, scornful, oppressive, disrespectful,
depersonalizing, reactionary
rude, invasive, crass, dehumanizing, belittling,
crude, marginalizing, scurrilous, repugnant,
pejorative
critical, traditionalist, androcentric,
anti-egalitarian, categorical, generalizing,
biased, patriarchal, depersonalizing,
stereotypical
misogynistic, predatory, anti-egalitarian,
antagonistic, rape-advocating,
appearance-driven, assaultive, sexualizing,
disdainful, exploitative
chauvinistic, discriminatory, patriarchal,
misogynistic, dismissive, dehumanizing,
intolerant, androcentric, bigoted, inferiorizing
victim-blaming, inferiorizing, victimizing,
reproachful, belligerent, dismissive,
generalizing, patriarchal, homogenizing,
derogatory
generalizing, victim-blaming, victimizing,
homogenizing, inflammatory, patriarchal,
essentialist, dismissive, inferiorizing,
prejudiced
overgeneralizing, moralizing, reductive,
patronizing, homogenizing, biased, censuring,
categorical, condescending, dehumanizing
belligerent, body-focused, intolerant,
essentialist, critical, paternalistic, devaluing,
misogynistic, invasive, bigoted
commodifying, bigoted, domineering,
paternalistic, dehumanizing, diminishing,
victimizing, exploitative, patriarchal,
beauty-obsessed
invasive, rude, belligerent, pejorative,
ofensive, body-focused, crass, denigrating,
disdainful, rape-advocating
victimizing, belligerent, intolerant, bigoted,
ignorant, assaultive, disempowering,
rape-advocating, paternalistic, essentialist
IDEOLOGICAL - INEQUALITY
MISOGYNY NON-SEXUAL
VIOLENCE
@Mzavalagc tienes cara de mojigata, pero
detrás hay una mujer muy corrupta...</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Discussion &amp; Conclusion</title>
      <p>XLM-RoBERTa performs best in the binary classification task (Subtask 1.1), especially when
finetuned with additional topic-related data. Our SCBMT model, however, proves more suitable for
finegrained multilabel classification problems by outperforming fine-tuned transformers. This outcome
Global explanations for multi-label and multiclass classification tasks.
can be directly attributed to the complementary semantic information captured by our adjective-based
representation. Another advantage of our SCBMT model over the plain transformer approach is that the
transformer component in SCBMT is not fine-tuned, ofering a lightweight alternative to the
resourceintensive fine-tuning typically required by transformer models. In addition, both the SCBM and the
SCBMT model provide high interpretability.</p>
      <p>Interestingly, although our adjectives and prompts were only designed for English, SCBMT still
performed comparably well on Spanish data. In some cases, SCBMT even achieved better performance
on the Spanish data. This suggests a high degree of language robustness of our method. This robustness
may be attributed not only to the multilingual capabilities of large language models, but also to the
generalizability of the semantic cues captured by our adjective-based concepts. A promising direction for
future work involves employing more sophisticated prompting techniques to derive the adjective-based
representation. Furthermore, collaboration with domain experts could support the expansion of the
automatically generated adjective lexicon in order to further improve its coverage and interpretability
for the intended users.</p>
    </sec>
    <sec id="sec-5">
      <title>Acknowledgments</title>
      <p>This work has been funded by the Vienna Science and Technology Fund (WWTF) [10.47379/ICT20016]
and by the Austrian Research Promotion Agency FFG under project no. 898085 (project TrustAI) and
FO999904624 (project FairAI).</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used ChatGPT in order to: Grammar and spelling
check. After using this service, the authors reviewed and edited the content as needed and take full
responsibility for the publication’s content.
Provide me with 50 adjectives that can be used to describe and distinguish classes in a task that aims
to categorize the message according to the intention of the author. It is a ternary classification task: (i)
direct sexist message, (ii) reported sexist message and (iii) judgmental message. This distinction allows to
diferentiate sexism that is actually taking place in online platforms from sexism which is being sufered by
women in other situations but that is being reported in social networks with the aim of complaining and
ifghting against sexism. The following classes are defined:
• Direct sexist message: the intention was to write a message that is sexist by itself or incites to be
sexist, as in: “A woman needs love, to fill the fridge, if a man can give this to her in return for her
services (housework, cooking, etc), I don’t see what else she needs."
• Reported sexist message: the intention is to report and share a sexist situation sufered by a woman
or women in first or third person, as in: “Today, one of my year 1 class pupils could not believe he’d
lost a race against a girl."
• Judgmental message: the intention is judgmental, since the tweet describes sexist situations or
behaviors with the aim of condemning them, as in: “21st century and we are still earning 25% less
than men #Idonotrenounce."
A.2. Task 1.3
Many facets of a woman’s life may be the focus of sexist attitudes including domestic and parenting roles,
career opportunities, sexual image, and life expectations, to name a few. Provide me with 50 adjectives that
can be used to describe and distinguish following classes:
• Ideological and inequality: this category includes tweets that discredit the feminist movement in
order to devalue and defame the struggle of women in any aspect of their lives. It also includes tweets
that reject inequality between men and women, or present men as victims of gender-based oppression.
As in: “Think the whole equality thing is getting out of hand. We are diferent, that’s how we’re
made!"
• Stereotyping and dominance: this category includes messages that express false ideas about women
that suggest they are more suitable or inappropriate for certain tasks. It also includes any claim that
implies that men are somehow superior to women. As in: “Most women no longer have the desire or
the knowledge to develop a high quality character, even if they wanted to."
• Objectification: Objectification and physical stereotyping include messages where women are
presented as objects apart from their dignity and personal aspects. We also include messages that assume
or describe certain physical qualities that women must have in order to fulfill traditional gender
roles, for example, ideas that suggest that women should maintain a standard and ideal of beauty or
attacks on a woman’s body. As in: “No ofense but I’ve never seen an attractive African American
hooker. Not a single one."
• Sexual violence: this category includes messages where sexual suggestions, requests or harassment of
a sexual nature (rape or sexual assault) are made, as in: “I wanna touch your tits..you can’t imagine
what I can do on your body."
• Misogyny and non-sexual violence: this category includes expressions of hatred and violence towards
women, as in: “Domestic abuse is never okay... Unless your wife is a bitch."</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>C.</given-names>
            <surname>Demus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Pitz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Schütz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Probol</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Siegel</surname>
          </string-name>
          , D. Labudde,
          <article-title>DeTox: A comprehensive dataset for German ofensive language and conversation analysis</article-title>
          , in: K.
          <string-name>
            <surname>Narang</surname>
            ,
            <given-names>A. Mostafazadeh</given-names>
          </string-name>
          <string-name>
            <surname>Davani</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Mathias</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Vidgen</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          Talat (Eds.),
          <source>Proceedings of the 6th Workshop on Online Abuse and Harms (WOAH</source>
          <year>2022</year>
          ),
          <article-title>Association for Computational Linguistics</article-title>
          , Online,
          <year>2022</year>
          , pp.
          <fpage>54</fpage>
          -
          <lpage>61</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>M.</given-names>
            <surname>Schütz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Schindler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Siegel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Nazemi</surname>
          </string-name>
          ,
          <article-title>Automatic fake news detection with pre-trained transformer models</article-title>
          , in: A.
          <string-name>
            <surname>Del Bimbo</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Cucchiara</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Sclarof</surname>
            ,
            <given-names>G. M.</given-names>
          </string-name>
          <string-name>
            <surname>Farinella</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Mei</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Bertini</surname>
            ,
            <given-names>H. J.</given-names>
          </string-name>
          <string-name>
            <surname>Escalante</surname>
          </string-name>
          , R. Vezzani (Eds.),
          <source>Pattern Recognition. ICPR International Workshops and Challenges</source>
          , Springer International Publishing,
          <year>2021</year>
          , pp.
          <fpage>627</fpage>
          -
          <lpage>641</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>L.</given-names>
            <surname>Plaza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Carrillo-de-Albornoz</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Arcos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Spina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Amigó</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gonzalo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Morante</surname>
          </string-name>
          , Overview of EXIST 2025:
          <article-title>Learning with disagreement for sexism identification and characterization in tweets, memes, and TikTok videos. Experimental IR meets multilinguality, multimodality, and interaction</article-title>
          , in: J.
          <string-name>
            <surname>C. de Albornoz</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Gonzalo</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Plaza</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. G. S. de Herrera</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Mothe</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Piroi</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Rosso</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Spina</surname>
          </string-name>
          , G. Faggioli, N. Ferro (Eds.),
          <source>Proceedings of the 16th International Conference of the CLEF Association (CLEF), Lecture Notes in Computer Science</source>
          , Springer International Publishing,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>L.</given-names>
            <surname>Plaza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Carrillo-de-Albornoz</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Arcos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Spina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Amigó</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gonzalo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Morante</surname>
          </string-name>
          , Overview of EXIST 2025:
          <article-title>Learning with disagreement for sexism identification and characterization in tweets, memes, and TikTok videos (extended overview)</article-title>
          , in: G. Faggioli,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ferro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          , D. Spina (Eds.),
          <source>CLEF 2025 Working Notes, CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>L.</given-names>
            <surname>Plaza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Carrillo-de Albornoz</surname>
          </string-name>
          , E. Amigó,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gonzalo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Morante</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Spina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Chulvi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Maeso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Ruiz</surname>
          </string-name>
          ,
          <string-name>
            <surname>EXIST</surname>
          </string-name>
          <year>2024</year>
          :
          <article-title>sexism identification in social networks</article-title>
          , in: N.
          <string-name>
            <surname>Goharian</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Tonellotto</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>He</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Lipani</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>McDonald</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Macdonald</surname>
          </string-name>
          , I. Ounis (Eds.),
          <source>Proceedings of the 46th European Conference on Information Retrieval (ECIR)</source>
          , Springer-Verlag,
          <year>2024</year>
          , p.
          <fpage>498</fpage>
          -
          <lpage>504</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>F.</given-names>
            <surname>Rodríguez-Sánchez</surname>
          </string-name>
          , J. C. de Albornoz,
          <string-name>
            <given-names>L.</given-names>
            <surname>Plaza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mendieta-Aragón</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Marco-Remón</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Makeienko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Plaza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gonzalo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Spina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          ,
          <source>Overview of EXIST</source>
          <year>2022</year>
          :
          <article-title>sExism Identification in Social neTworks</article-title>
          ,
          <source>Procesamiento del Lenguaje Natural</source>
          <volume>69</volume>
          (
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7] OpenAI,
          <source>Introducing OpenAI o3 and o4-mini</source>
          ,
          <year>2025</year>
          . URL: https://openai.com/index/ introducing-o3
          <string-name>
            <surname>-</surname>
          </string-name>
          and
          <string-name>
            <surname>-</surname>
          </string-name>
          o4-mini/.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>L.</given-names>
            <surname>Plaza</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Carrillo-de Albornoz</surname>
          </string-name>
          , I. Arcos,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Spina</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Amigó</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gonzalo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Morante</surname>
          </string-name>
          ,
          <string-name>
            <surname>EXIST</surname>
          </string-name>
          <year>2025</year>
          :
          <article-title>Learning with disagreement for sexism identification and characterization in tweets, memes, and TikTok videos</article-title>
          ,
          <source>in: Proceedings of European Conference on Information Retrieval</source>
          , Springer,
          <year>2025</year>
          , pp.
          <fpage>442</fpage>
          -
          <lpage>449</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A.</given-names>
            <surname>Dubey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jauhri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Pandey</surname>
          </string-name>
          , et al.,
          <source>The Llama 3 herd of models, CoRR abs/2407</source>
          .21783 (
          <year>2024</year>
          ). arXiv:
          <volume>2407</volume>
          .
          <fpage>21783</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>M. AI</surname>
          </string-name>
          , LLaMA
          <volume>3</volume>
          .1 8B Instruct, https://huggingface.co/meta-llama
          <source>/Llama-3</source>
          .
          <fpage>1</fpage>
          <string-name>
            <surname>-</surname>
          </string-name>
          8B-Instruct,
          <year>2024</year>
          . Accessed:
          <fpage>2025</fpage>
          -06-19.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>F.</given-names>
            <surname>AI</surname>
          </string-name>
          ,
          <string-name>
            <surname>XLM-RoBERTa</surname>
            <given-names>Large</given-names>
          </string-name>
          , https://huggingface.co/FacebookAI/xlm-roberta-large,
          <year>2020</year>
          . Accessed:
          <fpage>2025</fpage>
          -06-19.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>A.</given-names>
            <surname>Conneau</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Khandelwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          , et al.,
          <article-title>Unsupervised Cross-lingual Representation Learning at Scale, in: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics (ACL</article-title>
          ),
          <source>Association for Computational Linguistics</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>8440</fpage>
          -
          <lpage>8451</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>