<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">UMUTeam at EXIST 2024: Multi-modal Identification and Categorization of Sexism by Feature Integration Notebook for the EXIST 2024 Lab at CLEF 2024</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Ronghao</forename><surname>Pan</surname></persName>
							<email>ronghao.pan@um.es</email>
							<affiliation key="aff0">
								<orgName type="department">Facultad de Informática</orgName>
								<orgName type="institution">Universidad de Murcia</orgName>
								<address>
									<addrLine>Campus de Espinardo</addrLine>
									<postCode>30100</postCode>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">José</forename><forename type="middle">Antonio</forename><surname>García-Díaz</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Facultad de Informática</orgName>
								<orgName type="institution">Universidad de Murcia</orgName>
								<address>
									<addrLine>Campus de Espinardo</addrLine>
									<postCode>30100</postCode>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Tomás</forename><surname>Bernal-Beltrán</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Facultad de Informática</orgName>
								<orgName type="institution">Universidad de Murcia</orgName>
								<address>
									<addrLine>Campus de Espinardo</addrLine>
									<postCode>30100</postCode>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Rafael</forename><surname>Valencia-García</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Facultad de Informática</orgName>
								<orgName type="institution">Universidad de Murcia</orgName>
								<address>
									<addrLine>Campus de Espinardo</addrLine>
									<postCode>30100</postCode>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">UMUTeam at EXIST 2024: Multi-modal Identification and Categorization of Sexism by Feature Integration Notebook for the EXIST 2024 Lab at CLEF 2024</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">712A42F381012E0BBFB0AA7C43DE1B4A</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:56+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Sexism identification</term>
					<term>sexism categorization</term>
					<term>source intention detection</term>
					<term>multi modal Feature Engineering</term>
					<term>natural language processing</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The fourth edition of the EXIST shared task is a multimodal identification and categorization of sexism. This edition will take place as a lab in CLEF 2024. The main innovation in this edition with respect 2023 is the addition of a multimodal task based on sexism identification and categorization with MEMEs. As before, this shared task compromises sexist documents written in English and Spanish. For the textual tasks, we rely on feature integration from some Large Language Models and linguistic features extracted from our custom tool. We rank 15th in Sexism Identification, 8th in Source Intention, and 20th in Sexism Categorization. For the multimodal task, we rely on the CLIP model to extract the embedded text and image, and then combine them by diagonal multiplication to obtain the classification models. We rank 33rd in Sexism Identification, and 18th in both Source Intention and Sexism Categorization.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Social networks have become central platforms for social activism and complaints, allowing movements like #MeToo, #8M, and #Time'sUp to spread rapidly. Through these channels, women around the world have shared their real-life experiences of abuse, discrimination, and other forms of sexism. However, Social networks also facilitate the spread of sexist, disrespectful, and hateful behavior. In this context, the development of automated tools is essential. These tools can help detect and warn against sexist behavior and discourse, estimate the frequency of sexist and abusive situations on social media, identify the most common forms of sexism, and analyze how sexism is expressed on these platforms. Traditional systems for detecting sexism frequently depend on predefined labels and fixed perspectives, which can miss the complexity and subjectivity of sexist statements. The challenge in identifying and addressing sexism stems from its inherently subjective assessment. In contrast, perspectivism offers a promising method for enhancing detection by incorporating diverse opinions and viewpoints. An important contribution to this field, which aims to address the problem of sexism identification within the paradigm of learning with disagreement, has been made by the project EXIST 2024: sEXism Identification in Social neTworks <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2,</ref><ref type="bibr" target="#b2">3]</ref>.</p><p>Here we describe our participation in the fourth edition of EXIST <ref type="bibr" target="#b3">[4,</ref><ref type="bibr" target="#b4">5]</ref>. In the first three editions, EXIST focused only on the detection and classification of sexist text messages, and this 2024 edition introduces tasks that focus on images, specifically memes. Memes, generally humorous images, spread rapidly through social networks and the Internet. They can therefore encompass a wider range of sexist manifestations on social networks, especially those disguised as humor. Therefore, this shared task involves the development of automated multimodal tools capable of detecting sexism in both text and memes. With this addition, the organizers aim to cover a broader spectrum of sexism on social networks, especially that which is disguised as humor.</p><p>As a reminder, the tasks in the last edition of EXIST were three challenges, namely, (1) sexism identification, a binary classification task in which participants had to determine whether a tweet was sexist or not; (2) source intention, a multi-classification task focused on determining whether the author's intention was to post a sexist message, to report a sexist situation, or to make a judgment; and (3) sexism categorization, a multi-label classification task focused on identifying sexist characteristics. It should be noted that in EXIST 2024, the organizers are retaining the learning with disagreements paradigm proposed in 2023.</p><p>Our research group has experience in the detection hate speech <ref type="bibr" target="#b5">[6]</ref> and misogyny <ref type="bibr" target="#b6">[7]</ref> through the compilation and evaluation of several corpora. We have focused mainly on Spanish-language data and our work dealing with English-language data is more limited, limited to participating in the previous editions of EXIST <ref type="bibr" target="#b7">[8,</ref><ref type="bibr" target="#b8">9,</ref><ref type="bibr" target="#b9">10]</ref> and the evaluation of zero and few-shot learning strategies on some existing English datasets focused on hate speech detection <ref type="bibr" target="#b10">[11]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related works</head><p>Sexism refers to any abuse or negative sentiment directed at women based on their gender, or their gender in combination with other identity attributes. In particular, sexism is a growing problem on the Internet, with detrimental effects on women and other marginalized groups. It makes online spaces less accessible and less welcoming, perpetuating asymmetries and social injustices.</p><p>Automated tools are already widely used to identify and rate sexist content online. Researchers have proposed several approaches to address this problem, ranging from rule-based methods <ref type="bibr" target="#b11">[12]</ref> to the use of more complex models based on deep learning and pre-trained language models with Transformer architecture from a linguistic perspective <ref type="bibr" target="#b12">[13,</ref><ref type="bibr" target="#b13">14]</ref>, with only a few attempts to address the problem from a visual or multimodal perspective. Another work is described in <ref type="bibr" target="#b14">[15]</ref>, where the authors study about 6 thousand misogyny memes using a deep learning model that determines which modality plays a more significant role. The dataset included various characteristics such as hate speech, sexism, or cyberbullying, among others. The authors found that all modalities were useful for identifying misogyny, with text playing a significant role.</p><p>Sexism can be further categorized into different forms depending on the author's intent or the type of sexism. All editions of EXIST use categories such as "Ideology and Inequality", "Stereotyping and Dominance", "Objectification", "Sexual Violence", and "Misogyny and Non-Sexual Violence". This is similar to SemEval 2023 -Task 10 -Explainable Detection of Online Sexism (EDOS) <ref type="bibr" target="#b15">[16]</ref>, which defined a taxonomy for the more explainable classification of sexism in three hierarchical levels: binary sexism detection, category of sexism, and fine-grained vector of sexism. In all editions of EXIST, sexism is also considered at three levels: binary sexism detection in tweets, multiclass detection for source intention in tweets, and category of sexism in tweets.</p><p>Although the problem of sexism is primarily approached from a textual perspective, a field of research has emerged that approaches the problem from a visual or multimodal perspective, as the multimodal approach has shown superior performance to single modality approaches in detecting hate speech and misogyny, such as <ref type="bibr" target="#b16">[17]</ref> and <ref type="bibr" target="#b17">[18]</ref>. Sexist content can also appear in images or in multimodal forms. From a visual perspective, most research focuses on detecting offensive, non-conforming, or pornographic content. For this reason, this edition of EXIST includes a new task focused on detecting sexism in memes.</p><p>Currently, one of the problems that needs to be addressed in sexism detection is the presence of biases that could affect the actual performance of the models. Studies such as, <ref type="bibr" target="#b18">[19]</ref>, and <ref type="bibr" target="#b19">[20]</ref> have addressed this issue by using textual data to detect sexism. However, many of the contributions of previous sexism detection studies have not considered the complexity and multiple perspectives surrounding sexism, as sexism can manifest itself in multiple ways influenced by cultural, social, and individual factors.</p><p>Perspectivism in sexism detection involves considering multiple cultural, social, and individual viewpoints and contexts when analyzing potentially sexist content. It is essential to better understand the different manifestations of sexism, avoid bias, and improve the accuracy of detection models. It promotes equity by recognizing the different ways in which sexism can manifest itself, thus contributing to systems that are fairer and more sensitive to cultural and social diversity. For example, <ref type="bibr" target="#b13">[14]</ref> examine the errors made by classification models and discuss the difficulty of automatically classifying sexism due to the subjectivity of labels and the complexity of the natural language used in social networks.</p><p>Thus, this edition of EXIST has taken a similar approach to the 2023 edition, adopting the Learning with Disagreement (LeWiDi) paradigm for both dataset development and system evaluation. Unlike traditional methods that rely on a single "correct" label for each example, LeWiDi trains models to process and learn from conflicting or diverse annotations. This allows the system to incorporate different annotator perspectives, biases, and interpretations, resulting in a more equitable learning process.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Dataset</head><p>The textual challenges on EXIST 2024 followed the same strategies as previous editions <ref type="bibr" target="#b20">[21,</ref><ref type="bibr" target="#b21">22]</ref>; that is, it includes tweets written in Spanish or English crawled with certain keywords that are commonly used to undervalue the role of women. In fact, for the textual dataset the dataset is the same as from 2023.</p><p>For the multimodal dataset, the organizers of EXIST 2024 defined a lexicon of 250 terms and phrases that lead to sexist memes, derived from those that have proven effective in identifying sexism in previous editions of EXIST. This set includes 112 terms in English and 138 in Spanish, covering diverse topics and including terms with varying degrees of use in both sexist and non-sexist contexts, all centered around women. These terms were used as search queries on Google Images to obtain the top 100 images per term. Rigorous manual cleaning procedures were applied to define memes and remove noise such as textless images, text-only images, advertisements, and duplicates. This resulted in a final set of over 3,000 memes per language. Given the heterogeneous proportion of memes per term, we discarded the most unbalanced seeds and ensured that each seed had at least five memes. The final dataset was curated to achieve a balanced distribution of memes per seed. To avoid selection bias, memes were randomly selected while maintaining the appropriate distribution per seed. This process resulted in over 2,000 memes per language for the training set and over 500 memes per language for the test set.</p><p>For the annotation process, the organizers considered two socio-demographic traits: gender and age range. Each meme was then annotated by six crowd sourced annotators selected through the Prolific app, following guidelines developed by two experts in gender issues. The authors also provided the annotators' education level, ethnicity, and country of residence. The idea is to reduce labeling bias that may arise from cultural differences among annotators.</p><p>In addition, following the Learning with Disagreements paradigm removes the assumption that items have a single, unambiguous interpretation in a given context. Therefore, the dataset does not have a single "gold" annotation, but participants can use the full range of annotations from all annotators. This allows us to capture the diversity and develop more robust systems. The organizers release all annotations from all annotators to the participants.</p><p>The custom validation split is created using stratification by label in an 80-20 ratio.</p><p>First, since we have all the annotations, we decided to train our models for Task 1 (sexism identification) as a regression task rather than a classification task; thus, a text is considered sexist if the regression model outputs a score greater than 2.5. The soft labels are the output of the regression model normalized to a range of 1-10. Note that this approach was also used by our team in the previous edition of EXIST. The hard labels of the textual modality can be found in Table <ref type="table" target="#tab_0">1</ref>.</p><p>Second, Table <ref type="table" target="#tab_1">2</ref> shows the label distribution for Tasks 2 and 5. The labels are: (1) direct, if the intent of the document is to write a message that is itself sexist or incites to be sexist; (2) judgmental, if the intent is to report and share a sexist situation; and (3) reported, if the intent was to judge.</p><p>Third, Table <ref type="table" target="#tab_2">3</ref> shows the label distribution for Tasks 3 and 6. The labels are: (1) ideological and inequality, if the message downplays feminism or, equality between men and women; (2) stereotyping and dominance, if the message contains stereotypes about social roles; and (3) objectification, if the message contains physical characteristics about beauty standards or hyper-sexualization. As can be seen, we extracted a split from the dataset provided for individual validation (in an 80-20 ratio) using label stratification. For Task 2, there is a significant imbalance between the labels, with direct sexism being the label with the most examples and judgments and reports with a similar number of examples. For Task 3, the dataset has more balance between the characteristics. We have included the number of unknown responses in the textual modality. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Methodology</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Textual modalities</head><p>For the Tasks 1, 2, and 3 we decided to keep our previous strategy of training both languages (Spanish and English) together in order to reduce the carbon footprint of training a model for each language but to reduce the number of models to 4. Since we are evaluating feature integration strategies, we decided to keep two Spanish LLMs, BETO and MarIA; and two multilingual LLMs, deBERTa v3 <ref type="bibr" target="#b22">[23]</ref> and XLMTwitter. In this sense, we removed from our pipeline the multilingual BERT, XLM, BERTIN, DistilBETO and ALBETO <ref type="bibr" target="#b23">[24]</ref>). We also extracted the linguistic features (LFs) using the UMUTextStats tool <ref type="bibr" target="#b24">[25]</ref>.</p><p>To fine-tune each LLM for each task, we performed hyperparameter tuning on 10 models. This involved testing different learning rates, varying the number of epochs from 1 to 5, experimenting with batch sizes of 8 and 16, and adjusting warm-up steps and weight decay to optimize the learning rate during the initial training phases. The results of this tuning process are detailed in the Table <ref type="table" target="#tab_3">4</ref>. Next, we extract the classification token for each tweet, LLM, and Task (1, 2, and 3) using Sentence-BERT <ref type="bibr" target="#b25">[26]</ref> to extract the [CLS] token. Now that each document is represented by a unique fixed-length vector, we can merge it with the LFs using different feature integration strategies. The first strategy, known as Knowledge Integration (KI), is to use the feature vectors to train a new multi-input neural network. The second strategy is based on Ensemble Learning (EL). For this, we train separate simple neural networks for each LLM and for the LFs, and combine the results by averaging probabilities and outputs.</p><p>The training of the KI and the individual models are shown in the Table <ref type="table" target="#tab_4">5</ref>. As expected, most architectures are shallow (one or two hidden layers), brick-shaped (same number of neurons in each layer). This is expected because the vectors already capture the meaning of the sentences and are tuned to the output tasks. However, the KI for Tasks 1 and 3 resulted in deep neural networks, but with few neurons in each layer, being a funnel shape in Task 1 and a triangle shape in Task 3. If we compare these results with those obtained in the previous edition, it draws our attention that for Task 3 we obtained the best results in 2023 with deep neural networks and complex shapes.</p><p>For the EL, we evaluate different strategies: (1) mode of predictions, (2) averaging of probabilities, and (3) obtaining the highest probability. The only exception is Task 1, where we only averaged the predictions of the model, since we considered it to be a regression task. Now we report our results with our custom validation split. For Task 1 (see Table <ref type="table" target="#tab_5">6</ref>), as we handled it as a regression task in order to account for the disagreement between the annotators. Therefore, we report using Explained Variance (EV), Root Mean Squared Logarithmic Error (RMSLE), Pearson R, R Square (R2), Mean Average Error (MAE), Mean Squared Error (MSE) and Root Mean Squared Error (RMSE). The best results were obtained using the KI strategy. For Tasks 2 and 3, the results with the custom validation split are shown in the Table <ref type="table" target="#tab_6">7</ref>. The models are scored and ranked using the macro average precision, recall, and F1-Score. Similar to the previous edition, our team achieved their best results with the custom validation split with the with KI for Task 2 with the best recall and F1-Score but with the best precision is obtained with an EL based on the mode. In the case of Task 3, the EL strategies achieve better results with an F1-Score of 53.085% averaging probabilities and, the best recall of 81.242% with the highest probability, but the best precision  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Multi modalities</head><p>Figure <ref type="figure" target="#fig_0">1</ref> shows the architecture used to implement the model for Task 4, Task 5 and Task 6. We can see that we have used the CLIP <ref type="bibr" target="#b26">[27]</ref> model (specifically, openai/clip-vit-base-patch32<ref type="foot" target="#foot_0">1</ref> ) to extract both image and text embeddings, and then multiply those embeddings using a diagonal operation to use them as input to a classification neural network to identify whether a meme is sexist or not. Diagonal multiplication refers to a specific mechanism used during the training process to effectively connect and align text and image representations. As shown in Figure <ref type="figure" target="#fig_0">1</ref>, during training with the visual input (meme) and the text input (textual content of the meme), the text embedding vector and the image embedding vector perform diagonal alignment. This means that the dot product between the text embedding vector and its corresponding image embedding vector is high when they are related (i.e., for correct pairs) and low when they are not related (i.e., for incorrect pairs). This is possible because the CLIP model trains and processes the text and images in the same shared feature space. CLIP is a model developed by OpenAI that can efficiently understand and associate images and text. This means that the model learns to correctly associate a descriptive text with its corresponding image and to distinguish it from other unrelated images. Each component of the architecture is described in detail below:</p><p>• Image Encoder (CLIP image encoder). The meme image is passed through the CLIP image encoder, which extracts a series of embeddings {I 1 ,I 2 ,I 3 ,. . . ,I N } These embeddings represent the visual properties of the image. • Text Encoder (CLIP text encoder). The text of the meme is fed into the CLIP text encoder, which produces a series of embeddings {T 1 ,T 2 ,T 3 ,. . . ,T N } These embeddings capture the semantic features of the text. • Diagonal multiplication. The image and text embeddings are combined using a diagonal multiplication operation. This involves multiplying each text embedding T i with the corresponding image embedding I i to create a new set of combined embedding {T 1 I 1 ,T 2 I 2 ,T 3 I 3 ,. . . ,T N I N }. • Classification head. The combined embeddings are passed through a classification neural network consisting of the following components:</p><p>-Dropout Layer. A layer with a dropout rate of 0.1 to prevent overfitting.</p><p>-Linear Layer. A linear layer that reduces the dimensionality to N features, where N is the combined embedding length. Activation Layer (ReLU): A ReLU activation layer to introduce nonlinearities. -Dropout Layer. Another layer with a dropout rate of 0.1 to prevent overfitting.</p><p>-Final Linear Layer. A linear layer that reduces the output to 2 neurons, corresponding to the output classes (sexist or non-sexist).</p><p>For Tasks 5 and 6 we also assigned the most repeated label of the annotators, and in case of a tie, female annotators have more weight. In addition, we have used the same approach of using the CLIP model to obtain embeddings of text and images and then, through a diagonal multiplication, to obtain a combined embedding that will be the input of a classification neural network.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Results</head><p>From an evaluation metrics perspective, organizers use two types of evaluation based on the learning with disagreement paradigm <ref type="bibr" target="#b27">[28]</ref>.</p><p>• Hard-Hard. This is a comparison between "hard" system outputs (final labels) and "hard" ground truth. The Information Contrast Measure (ICM) metric is used to measure the similarity to the ground truth categories. The F1-Score is also reported, although it is not ideal in this context because it does not take into account the relationships between labels. • Soft-Soft. This is a comparison between "soft" system outputs (probabilities) and "soft" ground truths. In this case, the ICM-soft metric (an extension of ICM) is used as the official metric.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Textual modality</head><p>We sent three runs for Tasks 1, 2, and 3 based on the results on the custom validation split. For Tasks 1, our runs are KI, EL, LFs. Our results for the Task 1 are described in Table <ref type="table" target="#tab_7">8</ref>. We ranked 15th, 19th, and 32nd for the soft-soft scheme for each run. As expected, our results were better in Spanish than English, explained by the usage of two Spanish LLMs, BETO and MarIA. The model based on LFs outperformed the baselines proposed.</p><p>The official results for Task 2 are shown in Table <ref type="table" target="#tab_8">9</ref>. The first run is based on KI, while the second and third runs are based on MarIA and multilingual DeBERTA. This is the task in which we obtained our best results, ranking 8th (7th in Spanish and 10th in English) with the KI strategy, totaling 38 submissions among all teams. It is worth noting that MarIA outperformed MDeBERTA in both Spanish and English.</p><p>Next, the results for Task 3 are reported in Table <ref type="table" target="#tab_9">10</ref>. In this case, we focus on the hard vs. hard scheme, as we do not calculate probabilities. The first run is based on EL averaging the probabilities, the second run is based on KI, and the third run is based on EL based on mode. We got our best results with the first run, except for the English results, where we got better results with EL based on mode. Since we compete in a Hard vs. Hard scheme, the ranking also includes the Macro F1-Score, which is 49.42% with the first run, 47.38% with the second run, and 48.21% with the third run.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">Multimodal</head><p>Table <ref type="table" target="#tab_0">11</ref> shows the official results for Task 4, which evaluates hard labels only. The UMUTeam consistently ranks around 33-35 in all evaluations, both ALL and language-specific such as English or Spanish. We have exceeded the two baselines suggested by the organizers, ranking 33rd with -0.2422 on ICM-Hard and 0.6963 on F1_YES (F1-Score of sexist labels).</p><p>Although the ICM-Hard metrics are negative, indicating space for improvement in similarity to ground truth, we present less negative values compared to the baselines. This means that our predictions are closer to the true labels. If we look only at the F1-Score of the model, our system would be in a better position. However, in terms of similarity to the ground truth (ICM-Hard), we have a lot of room for improvement. This may be due to the fact that our approach does not have a final softmax layer indicating the probability of each label, since we used the model logits directly for the prediction. It is also possible that it is due to a lack of additional adjustments and optimizations. In this case, we only evaluated a 2e-5 learning rate, a training batch size of 16, 20 epochs, and an epoch-based validation strategy.</p><p>Table <ref type="table" target="#tab_1">12</ref> provides the official results for Task 5, evaluating only hard labels. In summary, we have achieved 18th place in all evaluations, slightly behind the baseline majority class (17th place) and ahead Table <ref type="table" target="#tab_2">13</ref> shows the official results for Task 6, which is a multi-label classification problem where only hard labels are evaluated. Our system achieved several ranks depending on the scenario with a total of 36 submissions among all teams: 18th in Hard-Hard ALL, 21st in the first column of Hard-Hard ES, and 12th in the second column of Hard-Hard ES. In comparison, the model performs better than the baselines in most scenarios, especially in the second Hard-Hard ES set.</p><p>The ICM-Hard values obtained (-1.9511, -2.3853, -1.5817) are negative, but generally better than the baselines, indicating a higher similarity to the true labels. It also clearly outperforms the baseline minor class in all scenarios and performs better than the baseline majority class in terms of ICM-Hard.</p><p>With respect to Macro F1-Score, we obtained significantly higher values than the baselines, especially in the Hard-Hard ES evaluation, indicating a better balance between precision and recall and an overall balanced performance in all classes.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 13</head><p>Official results for Task 6, including only hard labels. We report the rank, the ICM-Hard metric and the F1-Score of the YES label between parentheses.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Hard-Hard ALL</head><p>Hard-Hard ES Hard-Hard ES </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusions and further work</head><p>This document describes UMUTEAM's proposal for EXIST 2024, which focuses on the identification and categorization of SEXISM in Spanish and English. This is a very interesting competition as it deals with the learning with disagreements paradigm, binary and multi-classification tasks as well as a new challenge based on multimodal features. For the textual tasks, we based our proposal on the integration of linguistic features with sentence embeddings extracted for four LLMs, including Spanish and multilingual variants. We achieved our best result, 8th place, in Task 2, the source detection task.</p><p>As in previous editions, we are satisfied with participation, since we achieve competitive results in all tasks, outperforming the proposed baselines. However, in this edition, we only consider the different annotator schemes in the binary classification tasks, since it is considered a regression problem, but we are satisfied that we have been able to participate in all multimodal tasks, incorporating image processing modeling tachniques into our pipeline.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: System architecture of Task 4.</figDesc><graphic coords="7,72.00,446.99,451.26,105.14" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Datasets statistics for Tasks 1 (left) and 4 (right) concerning sexism identification.</figDesc><table><row><cell></cell><cell></cell><cell>Task 1</cell><cell></cell><cell>Task 4</cell></row><row><cell>label</cell><cell>train</cell><cell cols="4">val total train val total</cell></row><row><cell cols="6">non-sexist 2306 1540 3846 2933 734 3667</cell></row><row><cell>sexist</cell><cell cols="2">2466 1646 4112</cell><cell>302</cell><cell>75</cell><cell>377</cell></row><row><cell>total</cell><cell cols="5">4772 3186 7958 3235 809 4044</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>Datasets statistics for Tasks 2 (left) and 5 (right) concerning source intention.</figDesc><table><row><cell></cell><cell></cell><cell>Task 2</cell><cell></cell><cell></cell><cell>Task 5</cell><cell></cell></row><row><cell>label</cell><cell>train</cell><cell cols="5">val total train val total</cell></row><row><cell>-</cell><cell cols="3">3175 2090 5265</cell><cell>-</cell><cell>-</cell><cell>-</cell></row><row><cell>direct</cell><cell>992</cell><cell cols="5">665 1657 2609 626 3261</cell></row><row><cell>judgmental</cell><cell>296</cell><cell>225</cell><cell>521</cell><cell cols="2">652 157</cell><cell>783</cell></row><row><cell>reported</cell><cell>309</cell><cell>206</cell><cell>515</cell><cell>-</cell><cell>-</cell><cell>-</cell></row><row><cell>total</cell><cell cols="6">4772 3186 7958 3261 783 4044</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc>Distribution of the datasets of the Tasks 3 (left) and 6 (right) concerning misogyny categorization.</figDesc><table><row><cell>Task 3</cell><cell>Task 6</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 4</head><label>4</label><figDesc>Hyperparameter optimization of LLMs.</figDesc><table><row><cell>LLM</cell><cell cols="5">learning rate epochs batch size warmup steps weight decay</cell></row><row><cell></cell><cell></cell><cell>Task 1</cell><cell></cell><cell></cell><cell></cell></row><row><cell>BETO</cell><cell>3.6e-05</cell><cell>4</cell><cell>16</cell><cell>0</cell><cell>0.044</cell></row><row><cell>MARIA</cell><cell>2.9e-05</cell><cell>3</cell><cell>16</cell><cell>250</cell><cell>0.21</cell></row><row><cell>MDEBERTA</cell><cell>3.1e-05</cell><cell>4</cell><cell>16</cell><cell>500</cell><cell>0.18</cell></row><row><cell>XLMTWITTER</cell><cell>4.7e-05</cell><cell>2</cell><cell>8</cell><cell>0</cell><cell>0.22</cell></row><row><cell></cell><cell></cell><cell>Task 2</cell><cell></cell><cell></cell><cell></cell></row><row><cell>BETO</cell><cell>3.8e-05</cell><cell>3</cell><cell>16</cell><cell>500</cell><cell>0.17</cell></row><row><cell>MARIA</cell><cell>3.8e-05</cell><cell>2</cell><cell>8</cell><cell>250</cell><cell>0.031</cell></row><row><cell>MDEBERTA</cell><cell>1.4e-05</cell><cell>5</cell><cell>8</cell><cell>500</cell><cell>0.018</cell></row><row><cell>XLMTWITTER</cell><cell>4.2e-05</cell><cell>5</cell><cell>8</cell><cell>1000</cell><cell>0.22</cell></row><row><cell></cell><cell></cell><cell>Task 3</cell><cell></cell><cell></cell><cell></cell></row><row><cell>BETO</cell><cell>2.6e-05</cell><cell>4</cell><cell>16</cell><cell>0</cell><cell>0.25</cell></row><row><cell>MARIA</cell><cell>4.5e-05</cell><cell>5</cell><cell>16</cell><cell>0</cell><cell>0.054</cell></row><row><cell>MDEBERTA</cell><cell>4.5e-05</cell><cell>5</cell><cell>16</cell><cell>500</cell><cell>0.29</cell></row><row><cell>XLMTWITTER</cell><cell>2.8e-05</cell><cell>4</cell><cell>8</cell><cell>500</cell><cell>0.17</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 5</head><label>5</label><figDesc>Results of the hyper-parameter optimization stage using Keras of the LFs (LF), each LLM and the multi-input neural network using Knowledge Integration (KI).</figDesc><table><row><cell>feature set</cell><cell>shape</cell><cell cols="3">layers neurons dropout</cell><cell cols="2">lr batch size activation</cell></row><row><cell></cell><cell></cell><cell></cell><cell>Task 1</cell><cell></cell><cell></cell><cell></cell></row><row><cell>LF</cell><cell>brick</cell><cell>2</cell><cell>64</cell><cell cols="2">0 0.001</cell><cell>32 linear</cell></row><row><cell>BETO</cell><cell>triangle</cell><cell>3</cell><cell>512</cell><cell cols="2">0.3 0.001</cell><cell>64 tanh</cell></row><row><cell>MARIA</cell><cell cols="2">long funnel 3</cell><cell>256</cell><cell>0.1</cell><cell>0.01</cell><cell>64 sigmoid</cell></row><row><cell>MDEBERTA</cell><cell>brick</cell><cell>4</cell><cell>16</cell><cell cols="2">0.2 0.001</cell><cell>32 selu</cell></row><row><cell cols="2">XLMTWITTER brick</cell><cell>2</cell><cell>8</cell><cell cols="2">0.3 0.001</cell><cell>32 sigmoid</cell></row><row><cell>KI</cell><cell>funnel</cell><cell>8</cell><cell>8</cell><cell cols="2">0.3 0.001</cell><cell>64 elu</cell></row><row><cell></cell><cell></cell><cell></cell><cell>Task 2</cell><cell></cell><cell></cell><cell></cell></row><row><cell>LF</cell><cell>brick</cell><cell>2</cell><cell>256</cell><cell>0</cell><cell>0.01</cell><cell>256 linear</cell></row><row><cell>BETO</cell><cell>brick</cell><cell>6</cell><cell>128</cell><cell>0</cell><cell>0.01</cell><cell>128 elu</cell></row><row><cell>MARIA</cell><cell>triangle</cell><cell>5</cell><cell>512</cell><cell>0</cell><cell>0.01</cell><cell>256 elu</cell></row><row><cell>MDEBERTA</cell><cell>brick</cell><cell>2</cell><cell>512</cell><cell>0</cell><cell>0.01</cell><cell>256 linear</cell></row><row><cell cols="2">XLMTWITTER brick</cell><cell>5</cell><cell>512</cell><cell>0</cell><cell>0.01</cell><cell>256 selu</cell></row><row><cell>KI</cell><cell>brick</cell><cell>3</cell><cell>128</cell><cell cols="2">0.2 0.001</cell><cell>512 sigmoid</cell></row><row><cell></cell><cell></cell><cell></cell><cell>Task 3</cell><cell></cell><cell></cell><cell></cell></row><row><cell>LF</cell><cell>brick</cell><cell>2</cell><cell>256</cell><cell>0</cell><cell>0.01</cell><cell>64 linear</cell></row><row><cell>BETO</cell><cell>brick</cell><cell>2</cell><cell>128</cell><cell cols="2">0 0.001</cell><cell>64 tanh</cell></row><row><cell>MARIA</cell><cell>brick</cell><cell>2</cell><cell>512</cell><cell>0</cell><cell>0.01</cell><cell>64 linear</cell></row><row><cell>MDEBERTA</cell><cell>brick</cell><cell>2</cell><cell>512</cell><cell>0</cell><cell>0.01</cell><cell>64 linear</cell></row><row><cell cols="2">XLMTWITTER brick</cell><cell>2</cell><cell>37</cell><cell cols="2">0 0.001</cell><cell>64 sigmoid</cell></row><row><cell>KI</cell><cell>triangle</cell><cell>7</cell><cell>16</cell><cell>False</cell><cell>0.01</cell><cell>32 selu</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 6</head><label>6</label><figDesc>Results with the custom validation split for Task 1, reported as a regression task. The results are organized by feature set. The first block is the linguistic features, the second block are the LLMs, and the third and fourth blocks are the Knowledge Integration (KI) strategy and an ensemble learning (EL) respectively.</figDesc><table><row><cell>feature-set</cell><cell cols="3">EV RMSLE PEARSONR</cell><cell>R2 MAE</cell><cell>MSE RMSE</cell></row><row><cell>LF</cell><cell>0.162</cell><cell>0.414</cell><cell cols="3">0.458 0.161 1.573 3.629</cell><cell>1.905</cell></row><row><cell>BETO</cell><cell>0.563</cell><cell>0.224</cell><cell cols="3">0.752 0.562 1.084 1.892</cell><cell>1.376</cell></row><row><cell>MARIA</cell><cell>0.574</cell><cell>0.218</cell><cell cols="3">0.757 0.574 1.080 1.843</cell><cell>1.358</cell></row><row><cell>MDEBERTA</cell><cell>0.575</cell><cell>0.216</cell><cell cols="3">0.758 0.575 1.082 1.837</cell><cell>1.356</cell></row><row><cell cols="2">XLMTWITTER 0.541</cell><cell>0.235</cell><cell cols="3">0.735 0.541 1.130 1.986</cell><cell>1.409</cell></row><row><cell>EL</cell><cell>0.584</cell><cell>0.227</cell><cell cols="3">0.770 0.584 1.100 1.801</cell><cell>1.342</cell></row><row><cell>KI</cell><cell>0.602</cell><cell>0.203</cell><cell cols="3">0.776 0.602 1.042 1.720 1.312</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 7</head><label>7</label><figDesc>Results with custom validation for Tasks 2 and 3. The results are organized with the LFs (LF), all LLMs separately, the Knowledge Integration strategy (KI), and the three evaluated ensemble learning strategies (EL). All metrics are macro weighted.</figDesc><table><row><cell></cell><cell></cell><cell>Task 2</cell><cell></cell><cell></cell><cell>Task 3</cell></row><row><cell>feature-set</cell><cell>precision</cell><cell cols="3">recall F1-Score precision</cell><cell>recall F1-Score</cell></row><row><cell>LF</cell><cell cols="2">38.788 32.025</cell><cell>31.999</cell><cell cols="2">36.837 57.090</cell><cell>42.527</cell></row><row><cell>BETO</cell><cell cols="2">52.360 42.170</cell><cell>43.607</cell><cell cols="2">53.073 48.725</cell><cell>50.538</cell></row><row><cell>MARIA</cell><cell cols="2">53.592 44.683</cell><cell>47.010</cell><cell cols="2">50.375 55.177</cell><cell>52.343</cell></row><row><cell>MDEBERTA</cell><cell cols="2">52.402 41.623</cell><cell>42.556</cell><cell cols="2">45.748 59.038</cell><cell>50.853</cell></row><row><cell>XLMTWITTER</cell><cell cols="2">52.636 43.477</cell><cell>45.960</cell><cell cols="2">52.535 50.167</cell><cell>50.969</cell></row><row><cell>KI</cell><cell cols="2">54.288 52.896</cell><cell>53.371</cell><cell cols="2">49.873 54.543</cell><cell>51.760</cell></row><row><cell>EL (HIGHEST)</cell><cell cols="2">53.906 40.815</cell><cell>42.943</cell><cell cols="2">33.973 81.242</cell><cell>47.466</cell></row><row><cell>EL (MEAN)</cell><cell cols="2">59.768 40.618</cell><cell>41.843</cell><cell cols="2">51.229 56.362</cell><cell>53.085</cell></row><row><cell>EL (MODE)</cell><cell cols="2">61.823 40.601</cell><cell>41.731</cell><cell cols="2">50.016 54.478</cell><cell>51.699</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_7"><head>Table 8</head><label>8</label><figDesc>Official results for Task 1 with the Soft vs Soft scheme, including the Spanish and English results. Ranking is by runs.</figDesc><table><row><cell></cell><cell cols="2">Soft vs Soft ALL</cell><cell cols="2">Soft vs Soft ES</cell><cell cols="2">Soft vs Soft EN</cell></row><row><cell>Team</cell><cell cols="6">Rank ICM-Soft Rank ICM-Soft Rank ICM-Soft</cell></row><row><cell>UMUTeam 1</cell><cell>15</cell><cell>0.6679</cell><cell>12</cell><cell>0.7510</cell><cell>16</cell><cell>0.5200</cell></row><row><cell>UMUTeam 2</cell><cell>19</cell><cell>0.50339</cell><cell>18</cell><cell>0.5720</cell><cell>20</cell><cell>0.3745</cell></row><row><cell>UMUTeam 3</cell><cell>32</cell><cell>-0.4170</cell><cell>30</cell><cell>-0.3691</cell><cell>29</cell><cell>-0.5207</cell></row><row><cell>baseline majority class</cell><cell>36</cell><cell>-2.3585</cell><cell>36</cell><cell>-2.5421</cell><cell>36</cell><cell>-2.1991</cell></row><row><cell>baseline minority class</cell><cell>40</cell><cell>-3.0717</cell><cell>37</cell><cell>-2.5742</cell><cell>40</cell><cell>-3.8158</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_8"><head>Table 9</head><label>9</label><figDesc>Official results for Task 2 with the Soft vs Soft scheme, including the Spanish and English results. Ranking is by runs.</figDesc><table><row><cell></cell><cell cols="2">Soft vs Soft ALL</cell><cell cols="2">Soft vs Soft ES</cell><cell cols="2">Soft vs Soft EN</cell></row><row><cell>Team</cell><cell cols="6">Rank ICM-Soft Rank ICM-Soft Rank ICM-Soft</cell></row><row><cell>UMUTeam 1</cell><cell>8</cell><cell>-1.9569</cell><cell>7</cell><cell>-1.7667</cell><cell>10</cell><cell>-2.2517</cell></row><row><cell>UMUTeam 2</cell><cell>12</cell><cell>-2.0533</cell><cell>10</cell><cell>-1.8027</cell><cell>12</cell><cell>-2.4821</cell></row><row><cell>UMUTeam 3</cell><cell>19</cell><cell>-3.3189</cell><cell>19</cell><cell>-3.0432</cell><cell>20</cell><cell>-3.8940</cell></row><row><cell>baseline majority class</cell><cell>27</cell><cell>-5.4460</cell><cell>30</cell><cell>-5.6674</cell><cell>25</cell><cell>-5.2028</cell></row><row><cell>baseline minority class</cell><cell cols="2">35 -32.95527</cell><cell>35</cell><cell>-28.7093</cell><cell>35</cell><cell>-39.4948</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_9"><head>Table 10</head><label>10</label><figDesc>Official results for Task 3 with the Hard vs Hard scheme, including the Spanish and English results. Ranking is by runs.</figDesc><table><row><cell></cell><cell cols="2">Hard vs Hard ALL</cell><cell cols="2">Hard vs Hard ES</cell><cell cols="2">Hard vs Hard EN</cell></row><row><cell>Team</cell><cell cols="6">Rank ICM-Hard Rank ICM-Hard Rank ICM-Hard</cell></row><row><cell>UMUTeam 1</cell><cell>20</cell><cell>-0.733</cell><cell>24</cell><cell>-0.69597</cell><cell>21</cell><cell>-0.8012</cell></row><row><cell>UMUTeam 2</cell><cell>24</cell><cell>-0.8719</cell><cell>26</cell><cell>-0.8245</cell><cell>25</cell><cell>-0.9513</cell></row><row><cell>UMUTeam 3</cell><cell>22</cell><cell>-0.7901</cell><cell>25</cell><cell>-0.81457</cell><cell>20</cell><cell>-2.4821</cell></row><row><cell>baseline majority class</cell><cell>30</cell><cell>-1.5984</cell><cell>30</cell><cell>-1.7269</cell><cell>28</cell><cell>-1.4563</cell></row><row><cell>baseline minority class</cell><cell>34</cell><cell>-3.1295</cell><cell>34</cell><cell>-3.3196</cell><cell>33</cell><cell>-2.9279</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://huggingface.co/openai/clip-vit-base-patch32</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work has been supported by projects LaTe4PoliticES (PID2022-138099OB-I00) funded by MICI-U/AEI/10.13039/501100011033 and the European Regional Development Fund (ERDF)-a way of making Europe, LT-SWM (TED2021-131167B-I00) funded by MICIU/AEI/10.13039/ 501100011033 and by the European Union NextGenerationEU/PRTR, "Services based on language technologies for political microtargeting" (22252/PDC/23) funded by the Autonomous Community of the Region of Murcia through the Regional Support Program for the Transfer and Valorization of Knowledge and Scientific Entrepreneurship of the Seneca Foundation, Science and Technology Agency of the Region of Murcia. Mr. Ronghao Pan is supported by the Programa Investigo grant, funded by the Region of Murcia, the Spanish Ministry of Labour and Social Economy and the European Union -NextGenerationEU under the "Plan de Recuperación, Transformación y Resiliencia (PRTR)".</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 11</head><p>Official results for Task 4, including only hard labels. We report the rank, the ICM-Hard metric and the F1-Score of the YES label between parentheses.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Hard-Hard ALL</head><p>Hard-Hard ES Hard-Hard ES of the baseline minor class (21st place).</p><p>Regarding the ICM-Hard values of our system (-1.1486, -1.0605, -1.2148), they are negative and higher in magnitude compared to the baseline majority class, indicating a lower similarity to the true labels. However, they are significantly better than those of the baseline minor class (-2.0637, -2.0866, -2.0410).</p><p>In terms of F1_YES values, our model is superior to both baselines in all scenarios, especially in the second ES column (0.2805), indicating better precision and recall in the positive class.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 12</head><p>Official results for Task 5, including only hard labels. We report the rank, the ICM-Hard metric and the F1-Score of the YES label between parentheses.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Hard-Hard ALL</head><p>Hard-Hard ES Hard-Hard ES </p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Overview of exist 2021: sexism identification in social networks</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">R</forename><surname>-S. Y</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jorge</forename><surname>Carrillo-De-Albornoz Y</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Laura</forename><surname>Plaza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Julio</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Paolo</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Miriam</forename><surname>Comet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Trinidad</forename><surname>Donoso</surname></persName>
		</author>
		<ptr target="http://journal.sepln.org/sepln/ojs/ojs/index.php/pln/article/view/6389" />
	</analytic>
	<monogr>
		<title level="j">Procesamiento del Lenguaje Natural</title>
		<imprint>
			<biblScope unit="volume">67</biblScope>
			<biblScope unit="page" from="195" to="207" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Overview of exist 2022: sexism identification in social networks</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">R</forename><surname>-S. Y</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jorge</forename><surname>Carrillo-De-Albornoz Y</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Laura</forename><surname>Plaza Y Adrián Mendieta-Aragón Y</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Guillermo</forename><surname>Marco-Remón Y Maryna Makeienko Y María Plaza Y</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Julio</forename><surname>Gonzalo Y Damiano Spina Y</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Paolo</forename><surname>Rosso</surname></persName>
		</author>
		<ptr target="http://journal.sepln.org/sepln/ojs/ojs/index.php/pln/article/view/6443" />
	</analytic>
	<monogr>
		<title level="j">Procesamiento del Lenguaje Natural</title>
		<imprint>
			<biblScope unit="volume">69</biblScope>
			<biblScope unit="page" from="229" to="240" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Overview of exist 2023: sexism identification in social networks</title>
		<author>
			<persName><forename type="first">L</forename><surname>Plaza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Carrillo-De Albornoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Morante</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Amigó</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Spina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-031-28241-6_68</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of ECIR&apos;</title>
				<meeting>ECIR&apos;</meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="volume">23</biblScope>
			<biblScope unit="page" from="593" to="599" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Overview of exist 2024 -learning with disagreement for sexism identification and characterization in social networks and memes. experimental ir meets multilinguality, multimodality, and interaction</title>
		<author>
			<persName><forename type="first">L</forename><surname>Plaza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Carrillo-De Albornoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Ruiz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Maeso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chulvi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Amigó</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Morante</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Spina</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Fifteenth International Conference of the CLEF Association (CLEF 2024)</title>
				<meeting>the Fifteenth International Conference of the CLEF Association (CLEF 2024)</meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="316" to="342" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Overview of exist 2024 -learning with disagreement for sexism identification and characterization in social networks and memes (extended overview</title>
		<author>
			<persName><forename type="first">L</forename><surname>Plaza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Carrillo-De Albornoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Ruiz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Maeso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chulvi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Amigó</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Morante</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Spina</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2024-Conference and Labs of the Evaluation Forum</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Evaluating feature combination strategies for hate-speech detection in spanish using linguistic features and transformers</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>García-Díaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Jiménez-Zafra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>García-Cumbreras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Valencia-García</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Complex &amp; Intelligent Systems</title>
		<imprint>
			<biblScope unit="page" from="1" to="22" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Detecting misogyny in spanish tweets. an approach based on linguistics features and word embeddings</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>García-Díaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Cánovas-García</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Colomo-Palacios</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Valencia-García</surname></persName>
		</author>
		<idno>doi:j.future.2020.08.032</idno>
	</analytic>
	<monogr>
		<title level="j">Future Generation Computer Systems</title>
		<imprint>
			<biblScope unit="volume">114</biblScope>
			<biblScope unit="page" from="506" to="518" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Umuteam at exist 2021: Sexist language identification based on linguistic features and transformers in spanish and english</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>García-Díaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Colomo-Palacios</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Valencia-García</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="volume">2943</biblScope>
			<biblScope unit="page" from="512" to="521" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Umuteam at exist 2022: Knowledge integration and ensemble learning for multilingual sexism identification and categorization using linguistic features and transformers</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>García-Díaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Jiménez-Zafra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Colomo-Palacios</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Valencia-García</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Iberian Languages Evaluation Forum</title>
				<meeting>the Iberian Languages Evaluation Forum<address><addrLine>IberLEF</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022. 2022</date>
			<biblScope unit="volume">3202</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>García-Díaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Pan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Valencia-García</surname></persName>
		</author>
		<title level="m">Umuteam at exist 2023: Sexism identification and categorisation fine-tuning multilingual large language models</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Leveraging zero and few-shot learning for enhanced model generality in hate speech detection in spanish and english</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>García-Díaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Pan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Valencia-García</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Mathematics</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="page">5004</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Revisiting sexism detection using psychological scales and adversarial samples</title>
		<author>
			<persName><forename type="first">M</forename><surname>Samory</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kohne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Flöck</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Wagner</surname></persName>
		</author>
		<ptr target="https://api.semanticscholar.org/CorpusID:235303716" />
	</analytic>
	<monogr>
		<title level="m">International Conference on Web and Social Media</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">Sexism prediction in spanish and english tweets using monolingual and multilingual bert and ensemble models</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">F M</forename><surname>De Paula</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">F</forename><surname>Da Silva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">B</forename><surname>Schlicht</surname></persName>
		</author>
		<idno>ArXiv abs/2111.04551</idno>
		<ptr target="https://api.semanticscholar.org/CorpusID:237579065" />
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">Sexism identification in tweets and gabs using deep neural networks</title>
		<author>
			<persName><forename type="first">A</forename><surname>Kalra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zubiaga</surname></persName>
		</author>
		<idno>ArXiv abs/2111.03612</idno>
		<ptr target="https://api.semanticscholar.org/CorpusID:243832598" />
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Unveiling misogyny memes: A multimodal analysis of modality effects on identification</title>
		<author>
			<persName><forename type="first">S</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Naseem</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Razzak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Salim</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Companion Proceedings of the ACM on Web Conference 2024</title>
				<imprint>
			<date type="published" when="2024">2024</date>
			<biblScope unit="page" from="1864" to="1871" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">SemEval-2023 task 10: Explainable detection of online sexism</title>
		<author>
			<persName><forename type="first">H</forename><surname>Kirk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Yin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Vidgen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Röttger</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2023.semeval-1.305</idno>
		<ptr target="https://aclanthology.org/2023.semeval-1.305.doi:10.18653/v1/2023.semeval-1.305" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 17th International Workshop on Semantic Evaluation (SemEval-2023), Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Ojha</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Doğruöz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Tayyar</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Madabushi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><surname>Kumar</surname></persName>
		</editor>
		<editor>
			<persName><surname>Sartori</surname></persName>
		</editor>
		<meeting>the 17th International Workshop on Semantic Evaluation (SemEval-2023), Association for Computational Linguistics<address><addrLine>Toronto, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="2193" to="2210" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Multimodal hate speech detection via multi-scale visual kernels and knowledge distillation architecture</title>
		<author>
			<persName><forename type="first">A</forename><surname>Chhabra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">K</forename><surname>Vishwakarma</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.engappai.2023.106991</idno>
		<ptr target="https://doi.org/10.1016/j.engappai.2023.106991" />
	</analytic>
	<monogr>
		<title level="j">Engineering Applications of Artificial Intelligence</title>
		<imprint>
			<biblScope unit="volume">126</biblScope>
			<biblScope unit="page">106991</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Mistra: Misogyny detection through text-image fusion and representation analysis</title>
		<author>
			<persName><forename type="first">N</forename><surname>Jindal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">K</forename><surname>Kumaresan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Ponnusamy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Thavareesan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Rajiakodi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">R</forename><surname>Chakravarthi</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.nlp.2024.100073</idno>
		<ptr target="https://doi.org/10.1016/j.nlp.2024.100073" />
	</analytic>
	<monogr>
		<title level="j">Natural Language Processing Journal</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page">100073</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Detection of abusive language: the problem of biased datasets</title>
		<author>
			<persName><forename type="first">M</forename><surname>Wiegand</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ruppenhofer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Kleinbauer</surname></persName>
		</author>
		<ptr target="https://api.semanticscholar.org/CorpusID:174799974" />
	</analytic>
	<monogr>
		<title level="m">North American Chapter of the Association for Computational Linguistics</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Challenges in automated debiasing for toxic language detection</title>
		<author>
			<persName><forename type="first">X</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sap</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Swayamdipta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Choi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Smith</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2021.eacl-main.274</idno>
		<ptr target="https://aclanthology.org/2021.eacl-main.274.doi:10.18653/v1/2021.eacl-main.274" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">P</forename><surname>Merlo</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Tiedemann</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Tsarfaty</surname></persName>
		</editor>
		<meeting>the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="3143" to="3155" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Automatic classification of sexism in social networks: An empirical study on twitter data</title>
		<author>
			<persName><forename type="first">F</forename><surname>Rodríguez-Sánchez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Carrillo-De Albornoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Plaza</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="219563" to="219576" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Overview of exist 2021: Sexism identification in social networks</title>
		<author>
			<persName><forename type="first">F</forename><surname>Rodríguez-Sánchez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Carrillo-De Albornoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Plaza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Comet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Donoso</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Procesamiento del Lenguaje Natural</title>
		<imprint>
			<biblScope unit="volume">67</biblScope>
			<biblScope unit="page" from="195" to="207" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Chen</surname></persName>
		</author>
		<idno>CoRR abs/2111.09543</idno>
		<ptr target="https://arxiv.org/abs/2111.09543.arXiv:2111.09543" />
		<title level="m">DeBERTaV3: Improving DeBERTa using ELECTRA-style pre-training with gradient-disentangled embedding sharing</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">ALBETO and DistilBETO: Lightweight spanish language models</title>
		<author>
			<persName><forename type="first">J</forename><surname>Cañete</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Donoso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Bravo-Marquez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Carvallo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Araujo</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/2022.lrec-1.457" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Thirteenth Language Resources and Evaluation Conference, LREC 2022</title>
				<editor>
			<persName><forename type="first">N</forename><surname>Calzolari</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Béchet</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Blache</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">K</forename><surname>Choukri</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Cieri</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Declerck</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Goggi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Isahara</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Maegaard</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Mariani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Mazo</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Odijk</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Piperidis</surname></persName>
		</editor>
		<meeting>the Thirteenth Language Resources and Evaluation Conference, LREC 2022<address><addrLine>Marseille, France</addrLine></address></meeting>
		<imprint>
			<publisher>European Language Resources Association</publisher>
			<date type="published" when="2022-06-25">20-25 June 2022. 2022</date>
			<biblScope unit="page" from="4291" to="4298" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Umutextstats: A linguistic feature extraction tool for spanish</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>García-Díaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">J</forename><surname>Vivancos-Vicente</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Almela</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Valencia-García</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Thirteenth Language Resources and Evaluation Conference</title>
				<meeting>the Thirteenth Language Resources and Evaluation Conference</meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="6035" to="6044" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Sentence-bert: Sentence embeddings using siamese bert-networks</title>
		<author>
			<persName><forename type="first">N</forename><surname>Reimers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/D19-1410</idno>
		<ptr target="https://doi.org/10.18653/v1/D19-1410.doi:10.18653/v1/D19-1410" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP 2019</title>
				<editor>
			<persName><forename type="first">K</forename><surname>Inui</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Jiang</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">V</forename><surname>Ng</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">X</forename><surname>Wan</surname></persName>
		</editor>
		<meeting>the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP 2019<address><addrLine>Hong Kong, China</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">November 3-7, 2019. 2019</date>
			<biblScope unit="page" from="3980" to="3990" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Learning transferable visual models from natural language supervision</title>
		<author>
			<persName><forename type="first">A</forename><surname>Radford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">W</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Hallacy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ramesh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Goh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sastry</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Askell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Mishkin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Clark</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International conference on machine learning</title>
				<meeting><address><addrLine>PMLR</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="8748" to="8763" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">Evaluating extreme hierarchical multi-label classification</title>
		<author>
			<persName><forename type="first">E</forename><surname>Amigó</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Delgado</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2022.acl-long.399</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics</title>
		<title level="s">Long Papers</title>
		<meeting>the 60th Annual Meeting of the Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="5809" to="5819" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
