<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Detection of Sexism on Social Media with Multiple Simple Transformers</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Chirayu Jhakal</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Khushi Singal</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Manan Suri</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Divya Chaudhary</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Bijendra Kumar</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ian Gorton</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Khoury College of Computer Sciences, Northeastern University</institution>
          ,
          <addr-line>Seattle</addr-line>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Netaji Subhas University of Technology (NSUT)</institution>
          ,
          <addr-line>Dwarka Sector-3, Dwarka, Delhi, 110078</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>Social media platforms have become virtual communication channels, allowing users to voice their thoughts and opinions. However, this openness and features of anonymity have also given rise to the proliferation of harmful and ofensive content, including sexism. This research aims at proposing a methodology and explores the use of diferent simple transformers. Monolingual Simple Transformers such as BERT, RoBERTa[1], BERTweet, DistilBERT, XLNet were evaluated on the EXIST2023 shared task challenge at the IberLEF2023 dataset. It was observed that RoBERTa has given the best results among all other transformers. The proposed approach has great scope for the eficient detection of sexist content on social media, aiding in the development of efective content moderation systems.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Sexism Detection</kwd>
        <kwd>Simple Transformer Models</kwd>
        <kwd>Natural Language Processing</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Social media platforms have been revolutionary in the way people communicate and express
themselves in this digital age. These platforms have become a ubiquitous part of our daily lives,
enabling users to share thoughts and opinions and engage in social interactions. However,
along with the numerous benefits, the openness and features like anonymity and accessibility
of social media have also given rise to a concerning issue - the rapid increase of ofensive and
harmful content, including sexism.</p>
      <p>Sexism, defined as discrimination, stereotyping, or prejudice based on gender, continues to be a
pervasive problem in society. The presence of such content on social media not only perpetuates
harmful gender biases on possibly young and impressionable minds but also undermines the
inclusivity and safety of these online spaces. Consequently, there is a pressing need to develop
efective methods for detecting and mitigating sexist content on social media.</p>
      <p>
        In this work, we mainly focus on using Natural Language Processing techniques and
state-ofthe-art models for sexism detection, which aims to identify if a specific sentence contains sexist
content. The salient features of the proposed methodology are:
• The use of transformer-based language models, such as BERT (Bidirectional Encoder
Representations from Transformers)[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], RoBERTa (Robustly Optimized BERT)[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ][
        <xref ref-type="bibr" rid="ref1">1</xref>
        ],
and BERTweet, etc. These models, pre-trained on vast amounts of textual data, have
demonstrated exceptional capabilities in understanding and processing the complexities
of human language[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
• The methodology involves training and fine-tuning the transformer models on the
EX
      </p>
      <p>IST2023 dataset.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        There has been an increase in the work and interest in sexism detection as all social media
platforms want to limit ofensive content and make their platforms more inclusive. Examples
of such work are from the EXIST2022 edition: In [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], the use of Multilingual Models [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]and
Data Augmentation has been employed, and researchers have leveraged multilingual models to
detect sexist content across diferent languages[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. To overcome the lack of data in specific
languages, data augmentation techniques have been employed [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. In [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], the English
dataset and the Spanish dataset have been trained separately using a HuggingFace transformer
[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] uses the ensemble of 5 classification monolingual models and back-translation has
been used for augmenting the data; while [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] used datasets of diferent languages such as French,
German, and Italian and translated them into the pivot languages and BERTweet and BETO
have been used for English and Spanish respectively in a gradual unfreezing-discriminative
finetuning fashion [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. In [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], pre-processing methods like TFIDF (Term Frequency-Inverse
Document Frequency), Bag of words, and word2vec have been employed [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] and have seen the
use of basic algorithms like Naive Bayes, Support Vector Machines (SVM), and Linear Regression
to achieve remarkable results. In [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ], a framework for sexism detection on social media via
ByT5 and TabNet has been implemented.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Method</title>
      <sec id="sec-3-1">
        <title>3.1. Dataset</title>
        <p>The EXIST2023 dataset comprises posts from social media platforms like Twitter and Gab,
accompanied by annotations categorizing them into diferent types of sexism. The dataset is
divided into separate train and test partitions. The training set contains 6920 instances in both
English (3,260) and Spanish (3,660) languages. Meanwhile, the test set consists of 2,076 instances,
with 978 posts in English and 1098 posts in Spanish. Each data instance is labeled with binary
labels (for task 1) to determine whether it is classified as sexist or non-sexist.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Exploratory Analysis</title>
        <p>
          Before proceeding with the preprocessing steps, we performed an exploratory analysis of the
dataset[
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. This analysis involved gaining insights into the distribution of sexist content,
identifying common patterns, and understanding the characteristics of the data. We examined
the frequency of sexist posts, the prevalent forms of sexist language and expressions, and any
potential biases in the dataset.
        </p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Preprocessing</title>
        <p>
          The preprocessing began with the segregation of English data and Spanish data because both
languages are structurally and semantically quite diferent, so applying various language
models to the data would be easy and eficient [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ]. After cleaning the data various
preprocessing techniques were applied [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]:
• Removing web addresses from text.
• Removing emoticons from the text.
• Removing unrecognized characters, emojis, and stickers from text.
• Removing special characters.
• Removing repeating patterns like aaaaa, bbbbb, 00 etc.
        </p>
        <p>• Stemming words using Porter stemmer</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Pre-trained Transformer Models</title>
        <p>
          • RoBERTa: Robustly Optimized BERT (RoBERTa) [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ] is an optimized variant of BERT that
incorporates additional pre-training techniques. We fine-tuned the pre-trained RoBERTa
model on our dataset to specifically detect and classify instances of sexist content. In
English models, we applied RoBERTa after pre-processing and got decent results but in
the Spanish model, we got even better results [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ].
• BERT: Bidirectional Encoder Representations from Transformers (BERT) [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] is a widely
used transformer model that captures bidirectional contextual information. We fine-tuned
the pre-trained BERT model on our dataset to identify sexist content with improved
accuracy. BERT didn’t give satisfactory results on the English dataset but performed
better on the Spanish dataset [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ] [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ] [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ].
• Distill-BERT[19]: DistilBERT is a distilled version of the BERT model that ofers a lighter
and faster alternative while maintaining comparable performance. We fine-tuned the
pretrained DistilBERT model on our dataset to detect sexist content eficiently. DistilBERT
was applied only on the Spanish dataset and gave poor results [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ].
• BERTweet [20]: We fine-tuned the pre-trained BERTweet model on our dataset, specifically
designed to handle social media text. BERTweet employs a specialized vocabulary and
tokenization scheme tailored to social media language, enabling it to efectively detect
and classify instances of sexism in social media posts [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ] [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ].
• CamemBERT: CamemBERT is a transformer model specifically designed for the French
language. It is trained on large-scale French corpora and exhibits strong language
understanding capabilities. We fine-tuned the pre-trained CamemBERT model on our dataset
to accurately detect and categorize instances of sexism in French social media posts.
CamemBERT was only applied to the Spanish dataset and it gave the best results [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ]
[
          <xref ref-type="bibr" rid="ref6">6</xref>
          ].
• XLNet: XLNet is a transformer model that employs permutation-based training, allowing
it to capture bidirectional and context-aware representations efectively. We fine-tuned
the pre-trained XLNet model on our dataset to enhance the detection and classification
of sexist content. XlNet on the English dataset gave the best results among other models
[
          <xref ref-type="bibr" rid="ref7">7</xref>
          ] [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ] [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ].
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <p>In this study, we aimed to detect sexism in tweets written in both English and Spanish using
pre-trained transformer models. Specifically, we employed four diferent models for English
tweets, namely RoBERTa, BERT, BERTweet, and XLNet, while for Spanish tweets, we utilized
BERT, DistilBERT, RoBERTa, and CamemBERT. The performance of each model was evaluated
based on its accuracy in identifying instances of sexism in the tweets.</p>
      <p>For the English language, the RoBERTa model achieved an accuracy of 59.02% in detecting
sexism, while the BERT model achieved an accuracy of 56.16%. The BERTweet model, designed
specifically for tweets, achieved an accuracy of 69.60%, outperforming both RoBERTa and BERT.
The XLNet model, which incorporates a permutation-based approach, demonstrated the highest
accuracy among the English models, achieving 71.34%.</p>
      <p>In the case of Spanish tweets, the BERT model achieved an accuracy of 62.24% in detecting
sexism, followed by DistilBERT with an accuracy of 58.38%. The RoBERTa model exhibited an
accuracy of 62.77%, while the CamemBERT model demonstrated the highest accuracy among
the Spanish models, achieving 69.05%.</p>
      <p>These results indicate that the choice of pre-trained model has a significant impact on the
performance of sexism detection in tweets. While all models achieved relatively moderate
accuracy, the BERTweet model for English and the CamemBERT model for Spanish exhibited
the highest accuracies, suggesting their efectiveness in identifying instances of sexism in tweets.</p>
      <p>It is important to note that accuracy alone does not capture the full picture of model
performance, and other evaluation metrics, such as precision, recall, and F1 score, should be considered
for a comprehensive analysis. Furthermore, the generalizability of these models to diferent
datasets and domains should be further investigated to assess their robustness and applicability
in real-world scenarios.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Discussion</title>
      <p>The results of our study demonstrate the performance of various pre-trained transformer models
in detecting sexism in tweets written in both English and Spanish. Overall, the models exhibited
varying levels of accuracy, indicating their efectiveness in identifying instances of sexism in
social media content.</p>
      <p>
        For the English language, the BERTweet and XLNet models performed relatively better than
the RoBERTa[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] and BERT models[
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. This observation suggests that models specifically
designed for processing Twitter data, such as BERTweet, may be more suitable for capturing
the nuances and informal language commonly used in tweets. The XLNet model, which utilizes
a permutation-based approach, outperformed the other models, possibly due to its ability to
capture long-range dependencies in the text.
      </p>
      <p>In the case of Spanish tweets, the CamemBERT model displayed the highest accuracy among
the evaluated models. This indicates that CamemBERT, which is specifically trained on Spanish
text, is efective in capturing the linguistic characteristics and context-specific aspects of Spanish
tweets. However, it is worth noting that the accuracies achieved by the Spanish models were
relatively lower compared to the English models, suggesting the need for further research and
improvement in detecting sexism in Spanish-language tweets.</p>
      <p>It is important to consider the limitations of our study. First, the evaluation was performed
on a specific dataset, and the results may not be directly generalizable to other datasets or
real-world scenarios. Additionally, the performance of the models may vary depending on
the nature of the tweets, the distribution of sexism-related content, and cultural or contextual
factors. Further research is needed to assess the robustness and generalizability of these models
across diverse datasets and contexts.</p>
      <p>Moreover, accuracy alone may not be suficient to fully evaluate the performance of sexism
detection models. Additional metrics such as precision, recall, and F1 score should be considered
to assess the models’ ability to correctly identify instances of sexism while minimizing false
positives and false negatives.</p>
      <p>In conclusion, our study highlights the efectiveness of pre-trained transformer models in
detecting sexism in tweets, both in English and Spanish. The results demonstrate the importance
of using language-specific models and models designed for social media data to achieve higher
accuracy. This research contributes to the development of automated systems for identifying
and addressing sexism in online communication, ultimately fostering a more inclusive and
respectful digital environment.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>In this research, we investigated the detection of sexism in tweets using pre-trained transformer
models for both English and Spanish languages. Our results demonstrate that these models
can be efective in identifying instances of sexism in social media content. The BERTweet
model performed well in capturing the nuances of English tweets, while the CamemBERT
model showed promise for Spanish tweets. Additionally, the XLNet model exhibited superior
performance among the English models, highlighting the efectiveness of permutation-based
approaches. However, it is important to note that the accuracies achieved, especially for Spanish
models, can still be improved.</p>
      <p>The findings of this study have implications for developing automated systems that can
detect and mitigate sexism in online communication. By leveraging pre-trained transformer
models, we can gain insights into the prevalence of sexism and take steps toward fostering a
more inclusive and respectful digital environment.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Future Work</title>
      <p>While this research provides valuable insights into the detection of sexism in tweets, there are
several avenues for future work that can enhance the accuracy and robustness of the models.</p>
      <p>
        Firstly, data augmentation techniques can be employed to improve model performance [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]
[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. By increasing the diversity and quantity of training data through techniques such as
back-translation, word replacement, or text synthesis, we can potentially reduce the model’s
bias and enhance its ability to detect subtle forms of sexism.
      </p>
      <p>Secondly, ensemble modeling can be explored to leverage the strengths of multiple models
and improve overall performance. By combining predictions from diferent models, either by
majority voting or weighted averaging, we can potentially achieve higher accuracy and mitigate
the limitations of individual models.</p>
      <p>Furthermore, it is important to expand the evaluation of sexism detection models to diferent
languages and cultural contexts. The linguistic characteristics and contextual nuances can
significantly vary across languages, necessitating the development of language-specific models
and datasets.</p>
      <p>Additionally, further research should focus on addressing the issue of bias in the models. It is
crucial to identify and mitigate any biases encoded in the pre-trained models to ensure fair and
equitable detection of sexism.</p>
      <p>Finally, it would be beneficial to conduct user studies and assess the real-world impact of
automated systems in addressing sexism in online spaces. Understanding user perceptions,
reactions, and potential ethical concerns will guide the development of more efective and
responsible solutions.</p>
      <p>By pursuing these avenues, we can advance the field of sexism detection in social media and
contribute to the development of robust and inclusive technologies.
[19] A. Danday, T. S. Murthy, Twitter data analysis using distill bert and graph based convolution
neural network during disaster (2022).
[20] D. Q. Nguyen, T. Vu, A. T. Nguyen, Bertweet: A pre-trained language model for english
tweets, arXiv preprint arXiv:2005.10200 (2020).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ott</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Du</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Joshi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettlemoyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          ,
          <article-title>Roberta: A robustly optimized bert pretraining approach</article-title>
          , arXiv preprint arXiv:
          <year>1907</year>
          .
          <volume>11692</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>I.</given-names>
            <surname>Turc</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          ,
          <article-title>Well-read students learn better: On the importance of pre-training compact models</article-title>
          , arXiv preprint arXiv:
          <year>1908</year>
          .
          <volume>08962</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>D.</given-names>
            <surname>Liakhovets</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Schütz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Böck</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Andresel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kirchknopf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Babic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Slijpcevic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lampert</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Schindler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zeppelzauer</surname>
          </string-name>
          ,
          <article-title>Transfer learning for automatic sexism detection with multilingual transformer models</article-title>
          ,
          <source>in: Proceedings of the Iberian Languages Evaluation Forum (IberLEF</source>
          <year>2022</year>
          ).
          <article-title>CEUR-WS. org, A Coruna</article-title>
          ,
          <year>Spain</year>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>V.</given-names>
            <surname>Ahuir</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. Á.</given-names>
            <surname>González</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.-F.</given-names>
            <surname>Hurtado</surname>
          </string-name>
          ,
          <article-title>Enhancing sexism identification and categorization in low-data situations (</article-title>
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>R. L.</given-names>
            <surname>Tamayo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. O.</given-names>
            <surname>Bueno</surname>
          </string-name>
          ,
          <article-title>Are examples worth more than language? (</article-title>
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>D.</given-names>
            <surname>García-Baena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Á</surname>
          </string-name>
          .
          <string-name>
            <surname>García-Cumbreras</surname>
            ,
            <given-names>S. M.</given-names>
          </string-name>
          <string-name>
            <surname>Jiménez-Zafra</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>García-Vega</surname>
          </string-name>
          , Sinai at exist 2022:
          <article-title>Exploring data augmentation and machine translation for sexism identification (</article-title>
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaca-Serrano</surname>
          </string-name>
          ,
          <article-title>Detecting and classifying sexism by ensembling transformers models, language 2 (</article-title>
          <year>2022</year>
          )
          <article-title>1</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>V. P.</given-names>
            <surname>Álvarez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. M.</given-names>
            <surname>Vázquez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Chibane</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. L.</given-names>
            <surname>Domínguez</surname>
          </string-name>
          ,
          <article-title>Automatic sexism identification using an ensemble of pretrained transformers (</article-title>
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>F. M.</given-names>
            <surname>Plaza-del Arco</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.-D.</surname>
            Molina-González,
            <given-names>L. A.</given-names>
          </string-name>
          <string-name>
            <surname>Ureña-López</surname>
          </string-name>
          , M.-
          <string-name>
            <surname>T.</surname>
          </string-name>
          Martín-Valdivia,
          <article-title>Exploring the use of diferent linguistic phenomena for sexism identification in social networks (</article-title>
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>K.</given-names>
            <surname>Bengoetxea</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Aguirregoitia</surname>
          </string-name>
          , Multiaztertest@
          <fpage>exist</fpage>
          -iberlef2022:
          <article-title>Sexism identification in social networks (</article-title>
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A.</given-names>
            <surname>Rizvi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jamatia</surname>
          </string-name>
          ,
          <article-title>Nit-agartala-nlp-team at exist 2022: Sexism identification in social networks (</article-title>
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>A.</given-names>
            <surname>Younus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Qureshi</surname>
          </string-name>
          ,
          <article-title>A framework for sexism detection on social media via byt5 and tabnet (</article-title>
          <year>2022</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>J. W.</given-names>
            <surname>Tukey</surname>
          </string-name>
          , et al.,
          <article-title>Exploratory data analysis</article-title>
          , volume
          <volume>2</volume>
          , Reading, MA,
          <year>1977</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>H. T.</given-names>
            <surname>Ta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. B. S.</given-names>
            <surname>Rahman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Najjar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gelbukh</surname>
          </string-name>
          ,
          <article-title>Transfer learning from multilingual deberta for sexism identification</article-title>
          ,
          <source>in: CEUR Workshop Proceedings</source>
          , volume
          <volume>3202</volume>
          ,
          <string-name>
            <surname>CEUR-WS</surname>
          </string-name>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>A. F. M. de Paula</surname>
          </string-name>
          , R. F. da
          <string-name>
            <surname>Silva</surname>
          </string-name>
          ,
          <article-title>Detection and classification of sexism on social media using multiple languages, transformers, and ensemble models</article-title>
          ,
          <source>in: Proceedings of the Iberian Languages Evaluation Forum (IberLEF</source>
          <year>2022</year>
          )
          <article-title>co-located with the XXXVIII International Conference of the Spanish Society for Natural Language Processing (SEPLN</article-title>
          <year>2022</year>
          ), CEUR Workshop proceedings, volume
          <volume>3202</volume>
          ,
          <year>2022</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>11</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>W.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          , Bert:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          , arXiv preprint arXiv:
          <year>1810</year>
          .
          <volume>04805</volume>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>E.</given-names>
            <surname>Villa-Cueva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Sanchez-Vega</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. P.</given-names>
            <surname>López-Monroy</surname>
          </string-name>
          ,
          <article-title>Bi-ensembles of transformer for online bilingual sexism detection</article-title>
          ,
          <source>in: CEUR Workshop Proceedings</source>
          , volume
          <volume>3202</volume>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>G.</given-names>
            <surname>Shimi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mahibha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Thenmozhi</surname>
          </string-name>
          ,
          <article-title>Sexism identification in social media using deep learning models</article-title>
          ,
          <source>in: CEUR Workshop Proceedings</source>
          , volume
          <volume>3202</volume>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>