<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Sexism Identification in Social Networks: Advances in Automated Detection -A Report on the Exist Task at CLEF</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Nimra</forename><surname>Maqbool</surname></persName>
							<email>bsce21012@itu.edu.pk</email>
							<affiliation key="aff0">
								<orgName type="institution">Information Technology University (ITU)</orgName>
								<address>
									<settlement>Lahore</settlement>
									<country key="PK">Pakistan</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<address>
									<settlement>Grenoble</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Sexism Identification in Social Networks: Advances in Automated Detection -A Report on the Exist Task at CLEF</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">C61B93C1A690FB2EF5EDE082F547C956</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:54+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Sexism</term>
					<term>Gender discrimination</term>
					<term>Social networks</term>
					<term>Social media analysis</term>
					<term>Text classification</term>
					<term>Hate speech detection</term>
					<term>Natural language processing</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The widespread usage of social networks has brought forth numerous challenges, including the proliferation of sexist content, which contributes to gender inequality and discrimination. The Sexism Identification in Social Networks (SIT) task at the Conference and Labs of the Evaluation Forum (CLEF) aims to promote research and development of automated methods for identifying and mitigating sexism online. This review work provides an overview of the SIT task, discusses the dataset used, highlights the approaches and techniques employed, presents the evaluation metrics utilized, and offers insights into the future directions for advancing sexism identification in social networks. This review work describes the organization, goals, and results of the sExism Identification in Social networks (EXIST) challenge. EXIST 2024 proposes two challenges: sexism identification and sexism categorization of tweets and gabs, both in Spanish and English. During CLEF workshop, we investigate a broad range of models, including traditional machine learning methods, such as ensemble models and probability-based model like Random Forest, XGboost and deep learning architectures models, like BERT and Multilanguage models. The Experimental results show promising results in these areas and especially multilingual BERT model outperform among models.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>In the ever-evolving digital landscape, social networks have become an integral part of our daily lives. They provide us with platforms to connect, share ideas, and express ourselves. However, as these networks have expanded their reach, they have also become breeding grounds for various forms of discrimination and prejudice. Among these insidious phenomena, sexism, characterized by the marginalization and objectification of individuals based on their gender, remains a pervasive issue in online spaces.</p><p>Furthermore, Sexism in social networks presents a multifaceted challenge, as its manifestations range from subtle micro aggressions to overt harassment and abuse. Identifying and combating sexism in these virtual realms is crucial to fostering inclusive and equitable online environments. By doing so, we can not only promote gender equality but also cultivate safe and empowering digital spaces that resonate with users from diverse backgrounds.</p><p>The identification of sexism in social networks presents a formidable challenge due to the vast amounts of user-generated content, diverse linguistic expressions, and intricate nuances involved in language interpretation. Nevertheless, recent advances in artificial intelligence (AI) and natural language processing (NLP) have sparked renewed hope for combating this societal ill. By leveraging computational techniques and machine learning algorithms, researchers have begun to develop automated tools capable of detecting and analyzing instances of sexism within social network data. Within this study, I embark on an exploration of diverse machine learning and deep learning models, aiming to effectively classify labels pertaining to sexism identification, including both cross-grained labels and fine-grained labels.</p><p>The primary objective of this research endeavour revolves around three pivotal tasks, namely:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.1.">Task 1: Sexism Identification</head><p>In this task, the focus lies on binary classification. The underlying model is designed to discern whether a given tweet encompasses expressions or behaviors that exhibit sexism. This involves analyzing the content of the tweet to identify linguistic patterns, phrases, and contexts that may indicate sexist attitudes or discriminatory language.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.2.">Task 2: Source Intention</head><p>Once a message is identified as sexist, the subsequent task endeavours to categorize the message based on the author's intention. This classification sheds light on the role played by social networks in the generation and dissemination of sexist messages. The proposed task involves a ternary classification with the following categories:</p><p>1. Direct: The intention of the message is to be inherently sexist or to encourage sexist behaviour.</p><p>2. Reported: The intention is to report and share a sexist situation experienced by a woman, either in the first or third person. 3. Judgemental: The intention is to pass judgment, as the tweet describes sexist situations or behaviours with the aim of condemning them.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.3.">Task 3: Sexism Categorization</head><p>In this task, the objective is to further classify the tweets, specifically identifying the facets of women that are frequently targeted within social networks. This categorization will contribute to the development of policies aimed at combatting sexism. Each sexist tweet must be classified into one or more of the following categories:</p><p>1. Ideological and Inequality: The text discredits the feminist movement, denies the existence of gender inequality, or portrays men as victims of gender-based oppression. 2. Stereotyping and Dominance: The text perpetuates false notions about women, suggesting that they are more suited for certain roles (such as mother, wife, caregiver, affectionate, submissive, etc.) or unsuitable for certain tasks (such as driving, labour-intensive work, etc.). It may also claim male superiority over females. 3. Objectification: The text reduces women to objects, disregarding their dignity and personal qualities, or describes specific physical attributes that women must possess to conform to traditional gender roles. 4. Sexual Violence: The text contains sexual suggestions, requests for sexual favours, or engages in sexual harassment. 5. Misogyny and Non-Sexual Violence: The text expresses hatred and displays acts of violence towards women, which may not necessarily be of a sexual nature.</p><p>By comprehensively investigating these tasks, employing robust machine learning and deep learning models; we aim to enhance our understanding of sexism identification within social networks. This research will contribute to the development of more effective strategies for combatting sexism and fostering an inclusive and equitable online environment.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head><p>Sexism in social networks has emerged as a significant concern, highlighting the need for effective identification methods to mitigate its harmful effects. This section provides an overview of the existing research on sexism identification in social networks, including various approaches, methodologies, and key findings.</p><p>Francisco Rodr <ref type="bibr" target="#b0">[1]</ref> presenting a multilingual system based on pre-trained transformers and comparing single task to multi task learning to identify sexism in social networks. Similarly Rolfy Nixon Montufar Mercado <ref type="bibr" target="#b1">[2]</ref> proposed a model to detect cyberbullying in Spanish-language social networks using sentiment analysis techniques such as bag of words, elimination of signs and numbers, tokenization, stemming, and a Bayesian classifier. Dina Eliezer and Brenda Major <ref type="bibr" target="#b2">[3]</ref> examines whether group identification moderates the extent to which perceived in-group discrimination is threatening, as indexed by physiological and self-report measures. Women read and gave a speech summarizing an article describing sexism as prevalent or rare.</p><p>Furthermore, Simona Frenda <ref type="bibr" target="#b3">[4]</ref> propose first-time approach to the automatic detection of misogyny and sexism against women using the same computational approach. Along with that they also investigation of linguistic analogies and differences between sexism and misogyny from a computational point of view. He also examination of the usefulness of stylistic and lexical features for hate speech online against women. Parikh and Pulkit <ref type="bibr" target="#b4">[5]</ref> introduces a neural framework for multi-label classification of sexism and misogyny in texts, combining BERT with word embeddings, and outperforming existing baselines. Devadath <ref type="bibr" target="#b5">[6]</ref> examines Twitter sentiment analysis, focusing on NLP methods, tools, and machine learning algorithms like Naive Bayes. By evaluating metrics such as F1 score and precision, it categorizes a million tweets into positive and negative sentiments, emphasizing ethical considerations and informed decision-making.</p><p>Similarly, Devi and Bali <ref type="bibr" target="#b6">[7]</ref>introduces a refined classifier for identifying racist and sexist comments on Twitter using NLP and ML techniques. With XGBoost and word2vec, the model attained 69% accuracy and an F1 score of 0.690285, showcasing promising results amidst the pandemic-driven digital culture shift. Furthermore, De Paula, Angel Felipe Magnoss <ref type="bibr" target="#b7">[8]</ref> introduces a system for identifying and classifying sexism in English and Spanish social media using multilingual BERT models and ensemble strategies. It outperformed baseline models, securing first place in the EXIST 2021 task with high accuracies and F1-scores.</p><p>The EXIST campaign, focusing on online sexism detection, featured three tasks at CLEF 2023: sexism identification, categorization, and source intention identification, <ref type="bibr" target="#b8">[9]</ref>[10] as detailed by <ref type="bibr">Plaza et al. (2024)</ref>. Adopting a "learning with disagreement" approach, it aimed to reconcile differing perspectives in labeling, fostering equitable system development. The campaign's third edition, presented at the CLEF 2024 conference, provided new test and training data to enhance the identification and characterization of sexism in social networks and memes. With 28 participating teams and 232 submissions, the initiative underscored the research community's commitment to mitigating the impact of offensive content on women's well-being and freedom of expression on social media.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Utilization of Existing Models</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Machine Learning Models</head><p>Two machine learning models, Random Forest and XGBoost, were implemented using TF-IDF and BOW techniques for binary classification in Task 1 and multi-class classification in Task 2.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.1.">Random Forest</head><p>Random Forest, a popular ensemble learning technique, constructs multiple decision trees during training, using random subsets of data and features to reduce overfitting and enhance generalization. By combining predictions from these trees, it produces robust and accurate final predictions.</p><p>It was implemented using both TF-IDF and BOW techniques. Hyperparameter tuning was performed using RandomizedSearchCV to optimize the model, and it was subsequently trained using the best parameters obtained through this process. This approach was adopted to achieve the best possible results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.2.">XGBoost</head><p>XGBoost, or Extreme Gradient Boosting, is a fast and powerful implementation of gradient boosting algorithms. It sequentially builds decision trees to correct errors from previous trees, optimizing a specified loss function using gradient descent. Its efficiency and enhancements like regularization make it a popular choice for high-accuracy predictions in competitions and real-world applications.</p><p>The XGBoost model was implemented with both TF-IDF and BOW techniques. Hyperparameter tuning was performed using RandomizedSearchCV to identify the optimal parameters, and the model was trained using these best parameters. This approach aimed to achieve the highest performance.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Multi-language Bert Model</head><p>The Multilingual BERT (Bidirectional Encoder Representations from Transformers) model is a variant of the BERT model that is trained on text from multiple languages. BERT, developed by Google, is a pre-trained deep learning model that has achieved state-of-the-art results in various natural language processing (NLP) tasks.</p><p>The Multilingual BERT model is designed to handle multilingual text by jointly training on a large corpus of data from different languages. It learns a shared representation for words across languages, allowing it to capture cross-lingual semantic similarities and transfer knowledge between languages. This means that the model can understand and generate representations for text in multiple languages without needing language-specific models.</p><p>The training process of the Multilingual BERT model involves two key steps: pre-training and fine-tuning. During pre-training, the model is trained on a massive amount of monolingual text from various languages. It learns to predict missing words in sentences using a masked language modelling objective and also learns to determine whether two sentences follow each other in the original text or not, which helps it capture contextual relationships.</p><p>Once pre-training is completed, the model is fine-tuned on specific downstream NLP tasks such as text classification, named entity recognition, or sentiment analysis. Fine-tuning involves training the model on a smaller task-specific dataset in a supervised manner, where the model's pre-trained knowledge is adapted to the specific task at hand. This process allows the Multilingual BERT model to generalize across languages and perform well on various NLP tasks in different languages.</p><p>The Multilingual BERT model's success lies in its ability to capture contextual information from large-scale pre-training on diverse multilingual data. By leveraging the shared representations learned during pre-training, the model can handle a wide range of languages, even those with limited labelled data, and exhibit strong performance on tasks like text classification, information retrieval, and machine translation across multiple languages.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Experiments</head><p>The experiment section presents the findings and outcomes of the study on sexism identification in social networks. This section aims to provide a comprehensive analysis of the performance and effectiveness of the proposed approaches and methodologies. The results are presented in terms of various evaluation metrics, including accuracy, precision, recall, and F1-score. The section begins by summarizing the dataset used for evaluation, including the number of tweets, the languages covered, and the annotation process.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Dataset</head><p>Sexism encompasses any form of discrimination or bias against women based on their gender. In this study, we collected a large dataset of tweets in both English and Spanish, consisting of over 8 million tweets. The data collection period spanned from September 1, 2021, to September 30, 2022. To ensure a balanced dataset, we removed seeds (initial keywords) with fewer than 60 associated tweets. Ultimately, we obtained 183 seeds for Spanish and 163 seeds for English. The EXIST 2024 Tweets Dataset contains more than 10,000 labeled tweets, both in English and Spanish. In particular, the training set contains 6,920 tweets, the development set contains 1,038 tweets and the test set contains 2,076 tweets. Distribution between both languages has been balanced.</p><p>To address terminology and temporal biases, we carefully selected subsets of tweets for training, development, and testing. For each seed, we randomly selected approximately 20 tweets within the period from September 1, 2021, to February 28, 2022, for the training set. We aimed to maintain a representative temporal distribution within each seed's tweets. Similarly, we selected 3 tweets per seed from May 1, 2022, to May 31, 2022, for the development set, and 6 tweets per seed from August 1, 2022, to September 30, 2022, for the test set. To avoid author bias, we included only one tweet per author in the final selection. Additionally, we removed tweets containing fewer than 5 words. Consequently, we obtained over 3,200 tweets per language for the training set, around 500 per language for the development set, and nearly 1,000 tweets per language for the test set.</p><p>During the annotation process, we considered potential sources of "label bias. " Label bias may arise due to socio-demographic differences among annotators or when multiple correct labels or highly subjective decisions exist. To mitigate this bias, we took into account two demographic parameters: gender (MALE/FEMALE) and age (18-22 y.o./23-45 y.o./+46 y.o.). Each tweet was annotated by six crowdsourcing annotators selected through the Prolific app, following guidelines developed by two experts in gender issues.</p><p>Given the highly subjective nature of sexism identification and the challenge of interpreting natural language expressions in context, we adopted a learning with disagreements approach. This paradigm allows systems to learn from datasets where no definitive "gold" annotations are provided but instead incorporates information about the annotations from all six annotators, capturing the diversity of perspectives. By training directly from the data with disagreements, rather than relying on an aggregated label, we aim to incorporate the varying annotations per instance across the six different annotator strata.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Results</head><p>In this section, we present the findings and results obtained from our study. we employed various classification metrics to assess the performance of the sexism identification models. These metrics include accuracy, precision, recall, and F1-score, which provide a comprehensive evaluation of the models' ability to correctly classify tweets as sexist or non-sexist. In table <ref type="table" target="#tab_0">1</ref> the BERT model outperformed other models such as Random Forest and XGBoost in the task of binary classification of sexism in tweets, which involved both Spanish and English languages. BERT's success can be attributed to its advanced transformer architecture, pre-trained representations, multilingual capability, and ability to generate contextualized embeddings, allowing it to effectively capture complex linguistic patterns and understand context. In contrast, models like Random Forest and XGBoost with TF-IDF and BOW may have plateaued in performance due to their reliance on fixed-length feature representations and limitations in handling multilingual text. In table <ref type="table" target="#tab_1">2</ref> XGBoost and BERT outperformed the alternatives. XGBoost's strength lies in its ability to handle structured data and capture complex patterns through boosting, while BERT's deep contextual understanding of text allows it to excel in nuanced tasks. These models also generalize well to unseen data. In contrast, random forest models with TF-IDF and BOW underperformed, likely due to data imbalance, difficulty in handling high-dimensional sparse features, and lack of contextual understanding. Consequently, XGBoost and BERT demonstrated superior performance in our experiments.</p><p>In Task 2, the random forest model using TF-IDF struggled primarily due to data imbalance, high-dimensional sparse features, and a lack of contextual understanding. Imbalanced datasets skewed predictions towards the majority class, while TF-IDF's high-dimensional sparse vectors posed challenges for random forests not adept at handling such feature spaces. The model's inability to capture contextual relationships in text further undermined its performance. Solutions could involve resampling techniques to balance data classes, feature dimensionality reduction via PCA, or exploring alternative feature representations like word embeddings. Similarly, in Task 2 with Bag-of-Words (BOW), the random forest faced issues related to sparse feature handling and inadequate contextual comprehension. Improving performance might entail employing ensemble methods, incorporating more advanced feature engineering techniques, or integrating hybrid models that combine BOW with richer contextual information to better suit the task's requirements. In table 3, involving multi-label classification of the evaluation dataset using the multilingual BERT model, the evaluation metrics for all classes were in the 40s, indicating suboptimal performance. This likely resulted from the complexity of multi-label classification, data imbalance, and difficulty in capturing nuanced label distinctions. To improve performance, strategies such as data augmentation, balancing techniques, fine-tuning BERT, optimizing decision thresholds, model ensembling, and considering label dependencies can be applied. Additionally, post-processing methods like label smoothing can further refine accuracy. The table 4 include the Results given by the CLEF itself on test Dataset it include ICM-Hard which is a similarity function that generalizes Pointwise Mutual Information (PMI), and can be used to evaluate system outputs in classification problems by computing their similarity to the ground truth categories, ICM-hard Norm as well as F1-Score of Yes. The table 5 presents the performance rankings achieved by different models across three distinct tasks (Task-1, Task-2, and Task-3) as evaluated in the CLEF (Conference and Labs of the Evaluation Forum) runs. Each row specifies the type of model used (Random Forest, XGboost, Bert, Multilingual Bert) and the language dataset utilized (combined, Spanish, English). The ranks, ranging from 31 to 52, indicate how well each model performed relative to others within the same task and language context. Lower ranks signify better performance, showcasing which models and language configurations excelled in the evaluation scenarios.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusion</head><p>In conclusion, our study explored sexism classification using both traditional ensemble models (Random Forest and XGBoost) and a pretrained large language model (multilingual BERT). For the machine learning models, text encoding techniques such as TF-IDF and Bag of Words (BoW) were utilized. The multilingual BERT (mBERT) model outperformed all other models across the tasks, except for Task 2 where XGBoost demonstrated similar performance.</p><p>Despite the overall success of mBERT, the performance of all models was notably poorer in Task 2 and Task 3. This underperformance can be attributed to the highly imbalanced nature of the data, which posed significant challenges for model training and prediction accuracy. The imbalance likely resulted in a bias towards the majority class, reducing the models' ability to correctly classify the minority class instances. Nevertheless, this study provides valuable insights into the application of advanced machine learning techniques and pre-trained models for sexism classification in text.</p><p>Looking forward, research should expand comparisons among specialized large language models (LLMs) for English and Spanish to capitalize on language-specific nuances and improve classification accuracy. An effective strategy involves a language-specific approach where tweets are first categorized by language and then processed using dedicated pretrained models (e.g., English and Spanish). This method aims to enhance model performance by adapting analyses to each language's unique characteristics while maintaining overall classification integrity. While multilingual models like mBERT have performed well, future investigations could explore pretrained models exclusively trained on English and Spanish to mitigate challenges associated with multilingual training, such as nuanced language contexts and data biases. These specialized models are expected to enhance accuracy in sexism detection and other text classification tasks by better addressing specific linguistic nuances.</p><p>In summary, our study underscores the potential of advanced machine learning techniques and pretrained models in addressing sexism in social networks. Future research should prioritize refining model capabilities and addressing data challenges to enhance the effectiveness and applicability of sexism detection algorithms.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Performance metrics of TASK 1: Sexism Identification on Evaluation Dataset.</figDesc><table><row><cell>Model</cell><cell cols="4">Precision Recall F1-Score Accuracy</cell></row><row><cell>Random Forest (TF-IDF)</cell><cell>0.71</cell><cell>0.72</cell><cell>0.72</cell><cell>0.72</cell></row><row><cell>Random Forest (BOW)</cell><cell>0.73</cell><cell>0.72</cell><cell>0.71</cell><cell>0.72</cell></row><row><cell>XGboost (TF-IDF)</cell><cell>0.74</cell><cell>0.65</cell><cell>0.72</cell><cell>0.72</cell></row><row><cell>XGboost (BOW)</cell><cell>0.74</cell><cell>0.73</cell><cell>0.69</cell><cell>0.72</cell></row><row><cell>Multilingual Bert model</cell><cell>0.78</cell><cell>0.78</cell><cell>0.77</cell><cell>0.78</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>Performance metrics of TASK 2: Source Intention on Evaluation Dataset.</figDesc><table><row><cell>Model</cell><cell cols="4">Precision Recall F1-Score Accuracy</cell></row><row><cell>Random Forest (TF-IDF)</cell><cell>0.35</cell><cell>0.37</cell><cell>0.33</cell><cell>0.58</cell></row><row><cell>Random Forest (BOW)</cell><cell>0.37</cell><cell>0.34</cell><cell>0.31</cell><cell>0.55</cell></row><row><cell>XGboost (TF-IDF)</cell><cell>0.56</cell><cell>0.26</cell><cell>0.21</cell><cell>0.46</cell></row><row><cell>XGboost (BOW)</cell><cell>0.67</cell><cell>0.25</cell><cell>0.20</cell><cell>0.47</cell></row><row><cell>Multilingual Bert model</cell><cell>0.64</cell><cell>0.34</cell><cell>0.36</cell><cell>0.46</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc>Performance metrics of Multilingual Bert model for TASK 3: Source Intention on Evaluation Dataset.</figDesc><table><row><cell>Class Name</cell><cell cols="4">Precision Recall F1-Score Accuracy</cell></row><row><cell>Stereotyping-Dominance</cell><cell>0.42</cell><cell>0.41</cell><cell>0.39</cell><cell>0.40</cell></row><row><cell>Objectification</cell><cell>0.49</cell><cell>0.36</cell><cell>0.33</cell><cell>0.39</cell></row><row><cell>Sexual-Violence</cell><cell>0.44</cell><cell>0.46</cell><cell>0.42</cell><cell>0.44</cell></row><row><cell>Misogyny-Non-Sexual</cell><cell>0.41</cell><cell>0.36</cell><cell>0.37</cell><cell>0.38</cell></row><row><cell>Ideological-Inequality</cell><cell>0.34</cell><cell>0.33</cell><cell>0.38</cell><cell>0.35</cell></row><row><cell>Uknnown</cell><cell>0.0</cell><cell>0.0</cell><cell>0.0</cell><cell>0.0</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4</head><label>4</label><figDesc>Performance metrics of TASK 1, TASK 2 and TASK 3 on test dataset</figDesc><table><row><cell>Task</cell><cell>Model</cell><cell cols="4">Language ICM-Hard ICM-Hard Norm F1-YES</cell></row><row><cell>Task-1</cell><cell>Random Forest</cell><cell>Combine</cell><cell>0.28</cell><cell>0.64</cell><cell>0.69</cell></row><row><cell></cell><cell>Random Forest</cell><cell>ES</cell><cell>0.25</cell><cell>0.62</cell><cell>0.71</cell></row><row><cell></cell><cell>Random Forest</cell><cell>EN</cell><cell>0.29</cell><cell>0.64</cell><cell>0.66</cell></row><row><cell></cell><cell>XGboost</cell><cell>ES</cell><cell>0.33</cell><cell>0.66</cell><cell>0.71</cell></row><row><cell></cell><cell>XGboost</cell><cell>EN</cell><cell>0.34</cell><cell>0.67</cell><cell>0.68</cell></row><row><cell>Task-2</cell><cell>Random Forest</cell><cell>Combine</cell><cell>-0.47</cell><cell>0.34</cell><cell>0.29</cell></row><row><cell></cell><cell>Random Forest</cell><cell>ES</cell><cell>-0.51</cell><cell>0.33</cell><cell>0.29</cell></row><row><cell></cell><cell>Random Forest</cell><cell>EN</cell><cell>-0.42</cell><cell>0.35</cell><cell>0.30</cell></row><row><cell></cell><cell>XGboost</cell><cell>ES</cell><cell>-0.54</cell><cell>0.32</cell><cell>0.30</cell></row><row><cell></cell><cell>XGboost</cell><cell>Combine</cell><cell>-0.50</cell><cell>0.33</cell><cell>0.30</cell></row><row><cell></cell><cell>XGboost</cell><cell>EN</cell><cell>-0.46</cell><cell>0.33</cell><cell>0.29</cell></row><row><cell>Task-3</cell><cell>Bert</cell><cell>Combine</cell><cell>-2.34</cell><cell>0.00</cell><cell>0.17</cell></row><row><cell></cell><cell>Multilingual Bert model</cell><cell>ES</cell><cell>-2.29</cell><cell>0.00</cell><cell>0.17</cell></row><row><cell></cell><cell>Multilingual Bert model</cell><cell>EN</cell><cell>-2.40</cell><cell>0.00</cell><cell>0.15</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5</head><label>5</label><figDesc>Ranking achieved by runs of TASK 1, TASK 2, and TASK 3 on test dataset</figDesc><table><row><cell>Task</cell><cell>Model</cell><cell cols="2">Language Rank</cell></row><row><cell>Task-1</cell><cell>Random Forest</cell><cell>Both</cell><cell>51</cell></row><row><cell></cell><cell>Random Forest</cell><cell>ES</cell><cell>51</cell></row><row><cell></cell><cell>Random Forest</cell><cell>EN</cell><cell>52</cell></row><row><cell></cell><cell>XGboost</cell><cell>Both</cell><cell>45</cell></row><row><cell></cell><cell>XGboost</cell><cell>ES</cell><cell>45</cell></row><row><cell></cell><cell>XGboost</cell><cell>EN</cell><cell>47</cell></row><row><cell>Task-2</cell><cell>Random Forest</cell><cell>Both</cell><cell>35</cell></row><row><cell></cell><cell>Random Forest</cell><cell>ES</cell><cell>35</cell></row><row><cell></cell><cell>Random Forest</cell><cell>EN</cell><cell>32</cell></row><row><cell></cell><cell>XGboost</cell><cell>Both</cell><cell>36</cell></row><row><cell></cell><cell>XGboost</cell><cell>ES</cell><cell>36</cell></row><row><cell></cell><cell>XGboost</cell><cell>EN</cell><cell>34</cell></row><row><cell>Task-3</cell><cell>Bert</cell><cell>Both</cell><cell>33</cell></row><row><cell></cell><cell>Multilingual Bert</cell><cell>Es</cell><cell>33</cell></row><row><cell></cell><cell>Multilingual Bert model</cell><cell>EN</cell><cell>31</cell></row></table></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">A multi-task and multilingual model for sexism identification in social networks</title>
		<author>
			<persName><forename type="first">F</forename><surname>Rodríguez-Sánchez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">C</forename><surname>De Albornoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Plaza</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">IberLEF@SEPLN</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">Automatic cyberbullying detection in spanish-language social networks using sentiment analysis techniques</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">N</forename><surname>Mercado</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">The costs of caring: Gender identification increases threat following exposure to sexism</title>
		<author>
			<persName><forename type="first">D</forename><surname>Eliezer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Major</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">B</forename><surname>Mendes</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Experimental Social Psychology</title>
		<imprint>
			<biblScope unit="volume">46</biblScope>
			<biblScope unit="page" from="159" to="165" />
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Online hate speech against women: Automatic identification of misogyny and sexism on twitter</title>
		<author>
			<persName><forename type="first">S</forename><surname>Frenda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Ghanem</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Montes-Y Gómez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of intelligent &amp; fuzzy systems</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<biblScope unit="page" from="4743" to="4752" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Categorizing sexism and misogyny through neural approaches</title>
		<author>
			<persName><forename type="first">P</forename><surname>Parikh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Abburi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Chhaya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Varma</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Transactions on the Web (TWEB)</title>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="page" from="1" to="31" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">A comparison of multinomial naive bayes and xg boost for sentiment analysis and bias detection in tweets</title>
		<author>
			<persName><forename type="first">R</forename><surname>Devadath</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">C</forename><surname>Alex</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Sreeja</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Abhishek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Anjali</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">2024 International Conference on Wireless Communications Signal Processing and Networking (WiSPNET), IEEE</title>
				<imprint>
			<date type="published" when="2024">2024</date>
			<biblScope unit="page" from="1" to="7" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Racist tweets-based sentiment analysis using individual and ensemble classifiers</title>
		<author>
			<persName><forename type="first">B</forename><surname>Devi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">G</forename><surname>Shankar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Srivastava</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Nigam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Narang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Micro-Electronics and Telecommunication Engineering: Proceedings of 4th ICMETE 2020</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="555" to="567" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">F M</forename><surname>De Paula</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">F</forename><surname>Da Silva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">B</forename><surname>Schlicht</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2111.04551</idno>
		<title level="m">Sexism prediction in spanish and english tweets using monolingual and multilingual bert and ensemble models</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Overview of EXIST 2024 -Learning with Disagreement for Sexism Identification and Characterization in Social Networks and Memes</title>
		<author>
			<persName><forename type="first">L</forename><surname>Plaza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Carrillo-De-Albornoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Ruiz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Maeso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chulvi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Amigó</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Morante</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Spina</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fifteenth International Conference of the CLEF Association</title>
				<meeting><address><addrLine>CLEF</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2024">2024. 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Overview of EXIST 2024 -Learning with Disagreement for Sexism Identification and Characterization in Social Networks and Memes (Extended Overview)</title>
		<author>
			<persName><forename type="first">L</forename><surname>Plaza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Carrillo-De-Albornoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Ruiz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Maeso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chulvi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Amigó</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Morante</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Spina</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2024 -Conference and Labs of the Evaluation Forum</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">G S</forename><surname>Herrera</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
