<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Sexism Identification in Social Networks: Advances in Automated Detection - A Report on the Exist Task at CLEF</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Nimra Maqbool</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Information Technology University (ITU)</institution>
          ,
          <addr-line>Lahore</addr-line>
          ,
          <country country="PK">Pakistan</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The widespread usage of social networks has brought forth numerous challenges, including the proliferation of sexist content, which contributes to gender inequality and discrimination. The Sexism Identification in Social Networks (SIT) task at the Conference and Labs of the Evaluation Forum (CLEF) aims to promote research and development of automated methods for identifying and mitigating sexism online. This review work provides an overview of the SIT task, discusses the dataset used, highlights the approaches and techniques employed, presents the evaluation metrics utilized, and ofers insights into the future directions for advancing sexism identification in social networks. This review work describes the organization, goals, and results of the sExism Identification in Social networks (EXIST) challenge. EXIST 2024 proposes two challenges: sexism identification and sexism categorization of tweets and gabs, both in Spanish and English. During CLEF workshop, we investigate a broad range of models, including traditional machine learning methods, such as ensemble models and probability-based model like Random Forest, XGboost and deep learning architectures models, like BERT and Multilanguage models. The Experimental results show promising results in these areas and especially multilingual BERT model outperform among models.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Sexism</kwd>
        <kwd>Gender discrimination</kwd>
        <kwd>Social networks</kwd>
        <kwd>Social media analysis</kwd>
        <kwd>Text classification</kwd>
        <kwd>Hate speech detection</kwd>
        <kwd>Natural language processing</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>In the ever-evolving digital landscape, social networks have become an integral part of our daily
lives. They provide us with platforms to connect, share ideas, and express ourselves. However, as
these networks have expanded their reach, they have also become breeding grounds for various forms
of discrimination and prejudice. Among these insidious phenomena, sexism, characterized by the
marginalization and objectification of individuals based on their gender, remains a pervasive issue in
online spaces.</p>
      <p>Furthermore, Sexism in social networks presents a multifaceted challenge, as its manifestations range
from subtle micro aggressions to overt harassment and abuse. Identifying and combating sexism in
these virtual realms is crucial to fostering inclusive and equitable online environments. By doing so,
we can not only promote gender equality but also cultivate safe and empowering digital spaces that
resonate with users from diverse backgrounds.</p>
      <p>The identification of sexism in social networks presents a formidable challenge due to the vast
amounts of user-generated content, diverse linguistic expressions, and intricate nuances involved in
language interpretation. Nevertheless, recent advances in artificial intelligence (AI) and natural language
processing (NLP) have sparked renewed hope for combating this societal ill. By leveraging computational
techniques and machine learning algorithms, researchers have begun to develop automated tools capable
of detecting and analyzing instances of sexism within social network data. Within this study, I embark
on an exploration of diverse machine learning and deep learning models, aiming to efectively classify
labels pertaining to sexism identification, including both cross-grained labels and fine-grained labels.</p>
      <p>The primary objective of this research endeavour revolves around three pivotal tasks, namely:</p>
      <sec id="sec-1-1">
        <title>1.1. Task 1: Sexism Identification</title>
        <p>In this task, the focus lies on binary classification. The underlying model is designed to discern whether
a given tweet encompasses expressions or behaviors that exhibit sexism. This involves analyzing the
content of the tweet to identify linguistic patterns, phrases, and contexts that may indicate sexist
attitudes or discriminatory language.</p>
      </sec>
      <sec id="sec-1-2">
        <title>1.2. Task 2: Source Intention</title>
        <p>Once a message is identified as sexist, the subsequent task endeavours to categorize the message based
on the author’s intention. This classification sheds light on the role played by social networks in the
generation and dissemination of sexist messages. The proposed task involves a ternary classification
with the following categories:
1. Direct: The intention of the message is to be inherently sexist or to encourage sexist behaviour.
2. Reported: The intention is to report and share a sexist situation experienced by a woman, either
in the first or third person.
3. Judgemental: The intention is to pass judgment, as the tweet describes sexist situations or
behaviours with the aim of condemning them.</p>
      </sec>
      <sec id="sec-1-3">
        <title>1.3. Task 3: Sexism Categorization</title>
        <p>In this task, the objective is to further classify the tweets, specifically identifying the facets of women that
are frequently targeted within social networks. This categorization will contribute to the development
of policies aimed at combatting sexism. Each sexist tweet must be classified into one or more of the
following categories:
1. Ideological and Inequality: The text discredits the feminist movement, denies the existence of
gender inequality, or portrays men as victims of gender-based oppression.
2. Stereotyping and Dominance: The text perpetuates false notions about women, suggesting that
they are more suited for certain roles (such as mother, wife, caregiver, afectionate, submissive,
etc.) or unsuitable for certain tasks (such as driving, labour-intensive work, etc.). It may also
claim male superiority over females.
3. Objectification: The text reduces women to objects, disregarding their dignity and personal
qualities, or describes specific physical attributes that women must possess to conform to traditional
gender roles.
4. Sexual Violence: The text contains sexual suggestions, requests for sexual favours, or engages in
sexual harassment.
5. Misogyny and Non-Sexual Violence: The text expresses hatred and displays acts of violence
towards women, which may not necessarily be of a sexual nature.</p>
        <p>By comprehensively investigating these tasks, employing robust machine learning and deep learning
models; we aim to enhance our understanding of sexism identification within social networks. This
research will contribute to the development of more efective strategies for combatting sexism and
fostering an inclusive and equitable online environment.</p>
      </sec>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>Sexism in social networks has emerged as a significant concern, highlighting the need for efective
identification methods to mitigate its harmful efects. This section provides an overview of the existing
research on sexism identification in social networks, including various approaches, methodologies, and
key findings.</p>
      <p>
        Francisco Rodr [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] presenting a multilingual system based on pre-trained transformers and comparing
single task to multi task learning to identify sexism in social networks. Similarly Rolfy Nixon Montufar
Mercado [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] proposed a model to detect cyberbullying in Spanish-language social networks using
sentiment analysis techniques such as bag of words, elimination of signs and numbers, tokenization,
stemming, and a Bayesian classifier. Dina Eliezer and Brenda Major[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] examines whether group
identification moderates the extent to which perceived in-group discrimination is threatening, as
indexed by physiological and self-report measures. Women read and gave a speech summarizing an
article describing sexism as prevalent or rare.
      </p>
      <p>
        Furthermore, Simona Frenda [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] propose first-time approach to the automatic detection of misogyny
and sexism against women using the same computational approach. Along with that they also
investigation of linguistic analogies and diferences between sexism and misogyny from a computational point
of view. He also examination of the usefulness of stylistic and lexical features for hate speech online
against women. Parikh and Pulkit [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] introduces a neural framework for multi-label classification of
sexism and misogyny in texts, combining BERT with word embeddings, and outperforming existing
baselines. Devadath [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] examines Twitter sentiment analysis, focusing on NLP methods, tools, and
machine learning algorithms like Naive Bayes. By evaluating metrics such as F1 score and precision, it
categorizes a million tweets into positive and negative sentiments, emphasizing ethical considerations
and informed decision-making.
      </p>
      <p>
        Similarly, Devi and Bali [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]introduces a refined classifier for identifying racist and sexist comments
on Twitter using NLP and ML techniques. With XGBoost and word2vec, the model attained 69%
accuracy and an F1 score of 0.690285, showcasing promising results amidst the pandemic-driven digital
culture shift. Furthermore, De Paula, Angel Felipe Magnoss [8] introduces a system for identifying and
classifying sexism in English and Spanish social media using multilingual BERT models and ensemble
strategies. It outperformed baseline models, securing first place in the EXIST 2021 task with high
accuracies and F1-scores.
      </p>
      <p>The EXIST campaign, focusing on online sexism detection, featured three tasks at CLEF 2023: sexism
identification, categorization, and source intention identification,[ 9][10] as detailed by Plaza et al. (2024).
Adopting a "learning with disagreement" approach, it aimed to reconcile difering perspectives in
labeling, fostering equitable system development. The campaign’s third edition, presented at the CLEF
2024 conference, provided new test and training data to enhance the identification and characterization
of sexism in social networks and memes. With 28 participating teams and 232 submissions, the initiative
underscored the research community’s commitment to mitigating the impact of ofensive content on
women’s well-being and freedom of expression on social media.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Utilization of Existing Models</title>
      <sec id="sec-3-1">
        <title>3.1. Machine Learning Models</title>
        <p>Two machine learning models, Random Forest and XGBoost, were implemented using TF-IDF and BOW
techniques for binary classification in Task 1 and multi-class classification in Task 2.
3.1.1. Random Forest
Random Forest, a popular ensemble learning technique, constructs multiple decision trees during
training, using random subsets of data and features to reduce overfitting and enhance generalization.
By combining predictions from these trees, it produces robust and accurate final predictions.</p>
        <p>It was implemented using both TF-IDF and BOW techniques. Hyperparameter tuning was performed
using RandomizedSearchCV to optimize the model, and it was subsequently trained using the best
parameters obtained through this process. This approach was adopted to achieve the best possible
results.
3.1.2. XGBoost
XGBoost, or Extreme Gradient Boosting, is a fast and powerful implementation of gradient boosting
algorithms. It sequentially builds decision trees to correct errors from previous trees, optimizing a
specified loss function using gradient descent. Its eficiency and enhancements like regularization make
it a popular choice for high-accuracy predictions in competitions and real-world applications.</p>
        <p>The XGBoost model was implemented with both TF-IDF and BOW techniques. Hyperparameter
tuning was performed using RandomizedSearchCV to identify the optimal parameters, and the model
was trained using these best parameters. This approach aimed to achieve the highest performance.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Multi-language Bert Model</title>
        <p>The Multilingual BERT (Bidirectional Encoder Representations from Transformers) model is a variant
of the BERT model that is trained on text from multiple languages. BERT, developed by Google, is a
pre-trained deep learning model that has achieved state-of-the-art results in various natural language
processing (NLP) tasks.</p>
        <p>The Multilingual BERT model is designed to handle multilingual text by jointly training on a large
corpus of data from diferent languages. It learns a shared representation for words across languages,
allowing it to capture cross-lingual semantic similarities and transfer knowledge between languages.
This means that the model can understand and generate representations for text in multiple languages
without needing language-specific models.</p>
        <p>The training process of the Multilingual BERT model involves two key steps: pre-training and
ifne-tuning. During pre-training, the model is trained on a massive amount of monolingual text from
various languages. It learns to predict missing words in sentences using a masked language modelling
objective and also learns to determine whether two sentences follow each other in the original text or
not, which helps it capture contextual relationships.</p>
        <p>Once pre-training is completed, the model is fine-tuned on specific downstream NLP tasks
such as text classification, named entity recognition, or sentiment analysis. Fine-tuning involves
training the model on a smaller task-specific dataset in a supervised manner, where the model’s
pre-trained knowledge is adapted to the specific task at hand. This process allows the Multilingual
BERT model to generalize across languages and perform well on various NLP tasks in diferent languages.</p>
        <p>The Multilingual BERT model’s success lies in its ability to capture contextual information from
large-scale pre-training on diverse multilingual data. By leveraging the shared representations learned
during pre-training, the model can handle a wide range of languages, even those with limited labelled
data, and exhibit strong performance on tasks like text classification, information retrieval, and machine
translation across multiple languages.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments</title>
      <p>The experiment section presents the findings and outcomes of the study on sexism identification in social
networks. This section aims to provide a comprehensive analysis of the performance and efectiveness
of the proposed approaches and methodologies. The results are presented in terms of various evaluation
metrics, including accuracy, precision, recall, and F1-score. The section begins by summarizing the
dataset used for evaluation, including the number of tweets, the languages covered, and the annotation
process.</p>
      <sec id="sec-4-1">
        <title>4.1. Dataset</title>
        <p>Sexism encompasses any form of discrimination or bias against women based on their gender. In
this study, we collected a large dataset of tweets in both English and Spanish, consisting of over 8
million tweets. The data collection period spanned from September 1, 2021, to September 30, 2022. To
ensure a balanced dataset, we removed seeds (initial keywords) with fewer than 60 associated tweets.
Ultimately, we obtained 183 seeds for Spanish and 163 seeds for English. The EXIST 2024 Tweets
Dataset contains more than 10,000 labeled tweets, both in English and Spanish. In particular, the
training set contains 6,920 tweets, the development set contains 1,038 tweets and the test set contains
2,076 tweets. Distribution between both languages has been balanced.</p>
        <p>To address terminology and temporal biases, we carefully selected subsets of tweets for training,
development, and testing. For each seed, we randomly selected approximately 20 tweets within the
period from September 1, 2021, to February 28, 2022, for the training set. We aimed to maintain a
representative temporal distribution within each seed’s tweets. Similarly, we selected 3 tweets per seed
from May 1, 2022, to May 31, 2022, for the development set, and 6 tweets per seed from August 1, 2022,
to September 30, 2022, for the test set. To avoid author bias, we included only one tweet per author in
the final selection. Additionally, we removed tweets containing fewer than 5 words. Consequently,
we obtained over 3,200 tweets per language for the training set, around 500 per language for the
development set, and nearly 1,000 tweets per language for the test set.</p>
        <p>During the annotation process, we considered potential sources of "label bias." Label bias may arise
due to socio-demographic diferences among annotators or when multiple correct labels or highly
subjective decisions exist. To mitigate this bias, we took into account two demographic parameters:
gender (MALE/FEMALE) and age (18-22 y.o./23-45 y.o./+46 y.o.). Each tweet was annotated by six
crowdsourcing annotators selected through the Prolific app, following guidelines developed by two
experts in gender issues.</p>
        <p>Given the highly subjective nature of sexism identification and the challenge of interpreting natural
language expressions in context, we adopted a learning with disagreements approach. This paradigm
allows systems to learn from datasets where no definitive "gold" annotations are provided but instead
incorporates information about the annotations from all six annotators, capturing the diversity of
perspectives. By training directly from the data with disagreements, rather than relying on an aggregated
label, we aim to incorporate the varying annotations per instance across the six diferent annotator
strata.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Results</title>
        <p>In this section, we present the findings and results obtained from our study. we employed various
classification metrics to assess the performance of the sexism identification models. These metrics
include accuracy, precision, recall, and F1-score, which provide a comprehensive evaluation of the
models’ ability to correctly classify tweets as sexist or non-sexist.</p>
        <p>Precision</p>
        <p>Recall F1-Score</p>
        <p>Accuracy</p>
        <p>In table 1 the BERT model outperformed other models such as Random Forest and XGBoost
in the task of binary classification of sexism in tweets, which involved both Spanish and English
languages. BERT’s success can be attributed to its advanced transformer architecture, pre-trained
representations, multilingual capability, and ability to generate contextualized embeddings, allowing
it to efectively capture complex linguistic patterns and understand context. In contrast,
models like Random Forest and XGBoost with TF-IDF and BOW may have plateaued in performance
due to their reliance on fixed-length feature representations and limitations in handling multilingual text.</p>
        <p>In table 2 XGBoost and BERT outperformed the alternatives. XGBoost’s strength lies in its ability to
handle structured data and capture complex patterns through boosting, while BERT’s deep contextual
understanding of text allows it to excel in nuanced tasks. These models also generalize well to
unseen data. In contrast, random forest models with TF-IDF and BOW underperformed, likely due to
data imbalance, dificulty in handling high-dimensional sparse features, and lack of contextual
understanding. Consequently, XGBoost and BERT demonstrated superior performance in our experiments.</p>
        <p>In Task 2, the random forest model using TF-IDF struggled primarily due to data imbalance,
high-dimensional sparse features, and a lack of contextual understanding. Imbalanced datasets
skewed predictions towards the majority class, while TF-IDF’s high-dimensional sparse vectors
posed challenges for random forests not adept at handling such feature spaces. The model’s
inability to capture contextual relationships in text further undermined its performance. Solutions
could involve resampling techniques to balance data classes, feature dimensionality reduction via
PCA, or exploring alternative feature representations like word embeddings. Similarly, in Task
2 with Bag-of-Words (BOW), the random forest faced issues related to sparse feature handling
and inadequate contextual comprehension. Improving performance might entail employing
ensemble methods, incorporating more advanced feature engineering techniques, or integrating
hybrid models that combine BOW with richer contextual information to better suit the task’s requirements.</p>
        <p>In table 3, involving multi-label classification of the evaluation dataset using the multilingual BERT
model, the evaluation metrics for all classes were in the 40s, indicating suboptimal performance. This
likely resulted from the complexity of multi-label classification, data imbalance, and dificulty in
capturing nuanced label distinctions. To improve performance, strategies such as data augmentation,
balancing techniques, fine-tuning BERT, optimizing decision thresholds, model ensembling, and
considering label dependencies can be applied. Additionally, post-processing methods like label
smoothing can further refine accuracy.</p>
        <p>The table 4 include the Results given by the CLEF itself on test Dataset it include ICM-Hard which is
a similarity function that generalizes Pointwise Mutual Information (PMI), and can be used to evaluate
system outputs in classification problems by computing their similarity to the ground truth categories,
ICM-hard Norm as well as F1-Score of Yes.</p>
        <p>The table 5 presents the performance rankings achieved by diferent models across three distinct tasks
(Task-1, Task-2, and Task-3) as evaluated in the CLEF (Conference and Labs of the Evaluation Forum)
runs. Each row specifies the type of model used (Random Forest, XGboost, Bert, Multilingual Bert) and
the language dataset utilized (combined, Spanish, English). The ranks, ranging from 31 to 52, indicate
how well each model performed relative to others within the same task and language context. Lower
Model</p>
        <p>Language</p>
        <p>Rank
Task
Task-1
Task-2
Task-3</p>
        <p>Random Forest
Random Forest
Random Forest</p>
        <p>XGboost
XGboost</p>
        <p>XGboost
Random Forest
Random Forest
Random Forest</p>
        <p>XGboost
XGboost
XGboost</p>
        <p>Bert</p>
        <p>Multilingual Bert
Multilingual Bert model</p>
        <p>Both
ES
EN
Both
ES
EN
Both
ES
EN
Both
ES
EN
Both</p>
        <p>Es
EN
ranks signify better performance, showcasing which models and language configurations excelled in
the evaluation scenarios.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>In conclusion, our study explored sexism classification using both traditional ensemble models (Random
Forest and XGBoost) and a pretrained large language model (multilingual BERT). For the machine
learning models, text encoding techniques such as TF-IDF and Bag of Words (BoW) were utilized. The
multilingual BERT (mBERT) model outperformed all other models across the tasks, except for Task 2
where XGBoost demonstrated similar performance.</p>
      <p>Despite the overall success of mBERT, the performance of all models was notably poorer in Task 2
and Task 3. This underperformance can be attributed to the highly imbalanced nature of the data, which
posed significant challenges for model training and prediction accuracy. The imbalance likely resulted
in a bias towards the majority class, reducing the models’ ability to correctly classify the minority class
instances. Nevertheless, this study provides valuable insights into the application of advanced machine
learning techniques and pre-trained models for sexism classification in text.</p>
      <p>Looking forward, research should expand comparisons among specialized large language models
(LLMs) for English and Spanish to capitalize on language-specific nuances and improve classification
accuracy. An efective strategy involves a language-specific approach where tweets are first categorized
by language and then processed using dedicated pretrained models (e.g., English and Spanish). This
method aims to enhance model performance by adapting analyses to each language’s unique
characteristics while maintaining overall classification integrity. While multilingual models like mBERT have
performed well, future investigations could explore pretrained models exclusively trained on English
and Spanish to mitigate challenges associated with multilingual training, such as nuanced language
contexts and data biases. These specialized models are expected to enhance accuracy in sexism detection
and other text classification tasks by better addressing specific linguistic nuances.</p>
      <p>In summary, our study underscores the potential of advanced machine learning techniques and
pretrained models in addressing sexism in social networks. Future research should prioritize refining
model capabilities and addressing data challenges to enhance the efectiveness and applicability of
sexism detection algorithms.
[8] A. F. M. de Paula, R. F. da Silva, I. B. Schlicht, Sexism prediction in spanish and english tweets
using monolingual and multilingual bert and ensemble models, arXiv preprint arXiv:2111.04551
(2021).
[9] L. Plaza, J. Carrillo-de-Albornoz, V. Ruiz, A. Maeso, B. Chulvi, P. Rosso, E. Amigó, J. Gonzalo,
R. Morante, D. Spina, Overview of EXIST 2024 – Learning with Disagreement for Sexism
Identification and Characterization in Social Networks and Memes, in: Experimental IR Meets
Multilinguality, Multimodality, and Interaction. Proceedings of the Fifteenth International Conference of
the CLEF Association (CLEF 2024), 2024.
[10] L. Plaza, J. Carrillo-de-Albornoz, V. Ruiz, A. Maeso, B. Chulvi, P. Rosso, E. Amigó, J. Gonzalo,
R. Morante, D. Spina, Overview of EXIST 2024 – Learning with Disagreement for Sexism
Identification and Characterization in Social Networks and Memes (Extended Overview), in: G. Faggioli,
N. Ferro, P. Galuščáková, A. G. S. de Herrera (Eds.), Working Notes of CLEF 2024 – Conference
and Labs of the Evaluation Forum, 2024.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>F.</given-names>
            <surname>Rodríguez-Sánchez</surname>
          </string-name>
          , J. C. de Albornoz,
          <string-name>
            <given-names>L.</given-names>
            <surname>Plaza</surname>
          </string-name>
          ,
          <article-title>A multi-task and multilingual model for sexism identification in social networks</article-title>
          ,
          <source>in: IberLEF@SEPLN</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>R. N. Montufar</given-names>
            <surname>Mercado</surname>
          </string-name>
          ,
          <article-title>Automatic cyberbullying detection in spanish-language social networks using sentiment analysis techniques (</article-title>
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>D.</given-names>
            <surname>Eliezer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Major</surname>
          </string-name>
          , W. B.
          <string-name>
            <surname>Mendes</surname>
          </string-name>
          ,
          <article-title>The costs of caring: Gender identification increases threat following exposure to sexism</article-title>
          ,
          <source>Journal of Experimental Social Psychology</source>
          <volume>46</volume>
          (
          <year>2010</year>
          )
          <fpage>159</fpage>
          -
          <lpage>165</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Frenda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ghanem</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Montes-y Gómez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Rosso</surname>
          </string-name>
          ,
          <article-title>Online hate speech against women: Automatic identification of misogyny and sexism on twitter</article-title>
          ,
          <source>Journal of intelligent &amp; fuzzy systems 36</source>
          (
          <year>2019</year>
          )
          <fpage>4743</fpage>
          -
          <lpage>4752</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>P.</given-names>
            <surname>Parikh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Abburi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Chhaya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Varma</surname>
          </string-name>
          ,
          <article-title>Categorizing sexism and misogyny through neural approaches</article-title>
          ,
          <source>ACM Transactions on the Web (TWEB) 15</source>
          (
          <year>2021</year>
          )
          <fpage>1</fpage>
          -
          <lpage>31</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>R.</given-names>
            <surname>Devadath</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. C.</given-names>
            <surname>Alex</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Sreeja</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Abhishek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Anjali</surname>
          </string-name>
          ,
          <article-title>A comparison of multinomial naive bayes and xg boost for sentiment analysis and bias detection in tweets</article-title>
          ,
          <source>in: 2024 International Conference on Wireless Communications Signal Processing and Networking (WiSPNET)</source>
          , IEEE,
          <year>2024</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>7</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>B.</given-names>
            <surname>Devi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. G.</given-names>
            <surname>Shankar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Srivastava</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Nigam</surname>
          </string-name>
          , L. Narang,
          <article-title>Racist tweets-based sentiment analysis using individual and ensemble classifiers</article-title>
          ,
          <source>in: Micro-Electronics and Telecommunication Engineering: Proceedings of 4th ICMETE 2020</source>
          , Springer,
          <year>2021</year>
          , pp.
          <fpage>555</fpage>
          -
          <lpage>567</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>