<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">RMIT-IR at EXIST Lab at CLEF 2024 Notebook for the EXIST Lab at CLEF 2024</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Tony</forename><forename type="middle">Kim</forename><surname>Smith</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">RMIT University</orgName>
								<address>
									<settlement>Melbourne</settlement>
									<country key="AU">Australia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Ruda</forename><surname>Nie</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">Tay Nguyen University</orgName>
								<address>
									<settlement>Buon Ma Thuot</settlement>
									<country key="VN">Vietnam</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Johanne</forename><forename type="middle">R</forename><surname>Trippas</surname></persName>
							<email>j.trippas@rmit.edu.au</email>
							<affiliation key="aff0">
								<orgName type="institution">RMIT University</orgName>
								<address>
									<settlement>Melbourne</settlement>
									<country key="AU">Australia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Damiano</forename><surname>Spina</surname></persName>
							<email>damiano.spina@rmit.edu.au</email>
							<affiliation key="aff0">
								<orgName type="institution">RMIT University</orgName>
								<address>
									<settlement>Melbourne</settlement>
									<country key="AU">Australia</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">RMIT-IR at EXIST Lab at CLEF 2024 Notebook for the EXIST Lab at CLEF 2024</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">F5BE85C55E241D505993918B90A501BC</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:54+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>sexism characterization, large language models, in-context learning, multi-modal contrastive learning Orcid 0009-0009-4752-6679 (T. K. Smith)</term>
					<term>0000-0002-1194-2496 (H. R. Nie)</term>
					<term>0000-0002-7801-0239 (J. R. Trippas)</term>
					<term>0000-0001-9913-433X (D. Spina)</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper describes RMIT-IR team's participation in the EXIST Lab at CLEF 2024. The proposed approaches aim to address sexism characterization on microblog posts (Tasks 1, 2, and 3) and sexism identification on memes (Task 4). For Tasks 1-3, we studied the effectiveness of zero-shot In-Context Learning (ICL) <ref type="bibr" target="#b0">[1]</ref> with off-the-shelf pre-trained Large Language Models (LLMs) to mimic the scenario of minimal intervention of a practitioner aiming to build sexism characterization systems. Our approaches for meme classification (Task 4) utilize CLIP (Contrastive Language-Image Pre-training) [2] to experiment with multi-modal embeddings and zero-shot sexism identification models. We report the performance of our approaches under the learning with disagreements regime (Soft evaluation) and also for label predictions (Hard evaluation). The code of our submission is available at https://github.com/rmit-ir/exist2024/.</p><p>Warning: Some of the examples included in this paper may contain offensive language and explicit descriptions of sexist behavior, which may be disturbing to the reader.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Social media has had a considerable impact on human societies. Applications such as Facebook, YouTube, Instagram, and TikTok have helped move the zeitgeist while creating large communities in the millions. However, many social media platforms have issues with people creating and posting harmful information. The challenge of detecting and managing such harmful content remains a concern for social media companies, contributing to consequences ranging from misinformation to adverse effects on mental health <ref type="bibr" target="#b2">[3]</ref>. In addition, the rise of social media has empowered influencers who often unwittingly or deliberately propagate harmful stereotypes and negative gender norms. This type of content attracts an audience and drives advertising revenue, perpetuating a cycle of negativity <ref type="bibr" target="#b3">[4]</ref>. As a result, it often fosters negative behaviour towards women and minority groups, impacting many people negatively <ref type="bibr" target="#b4">[5]</ref>.</p><p>Sexism is the belief that the members of one sex or gender are less than the members of the other sex, especially that women are less able than men <ref type="bibr" target="#b5">[6]</ref>. This can be categorized into hostile sexism and benevolent sexism.</p><p>Sexism can limit the opportunities and roles people of different sexes and genders are expected to take. It can be conveyed through any form of expression, like images, cartoons, memes, objects, gestures, and symbols, and can be spread offline or online. This oppression can take different forms, such as economic exploitation and social domination <ref type="bibr" target="#b6">[7]</ref>.</p><p>Sexist attitudes and behaviours can perpetuate stereotypes of social and gender roles based on one's biological sex. Usually, people are socialized with sexist concepts that teach traditional gender roles for males and females <ref type="bibr" target="#b7">[8]</ref>. Hostile sexism represents a form of sexist ideology, marked by explicit hostility towards women and the perception of them as inferior and submissive <ref type="bibr" target="#b8">[9]</ref>. This deeply ingrained perception often results in the mistreatment of women at both individual and institutional levels <ref type="bibr" target="#b7">[8]</ref>. Benevolent sexism is a nuanced manifestation that ingrains in men the belief that they should be responsible for providing for women in intimate relationships <ref type="bibr" target="#b8">[9]</ref>. This belief system dictates specific roles and behaviours for women, such as expecting them to demonstrate motherly instincts, subtly reinforcing traditional gender roles. A society that has high rates of hostile and benevolent sexism often has high rates of violence against women, such as domestic violence, rape, and the commodification of women and their bodies <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b10">11]</ref>.</p><p>There has been a recent increase in research on identifying different forms of hate speech, corresponding with advancements in generative pre-trained transformers and, in general, large language models (LLMs) <ref type="bibr" target="#b11">[12]</ref>. Researchers are asking how LLMs can be trained to identify subtle and overt sexist content <ref type="bibr" target="#b12">[13,</ref><ref type="bibr" target="#b13">14,</ref><ref type="bibr" target="#b14">15]</ref>. However, many questions on how state-of-the-art LLMs can be used for sexism detection are still open. What criteria should be used to evaluate what constitutes sexism in varied cultural contexts? If a dataset with binary classifications<ref type="foot" target="#foot_0">1</ref> is employed, can a machine learning model accurately capture the nuances within the text? And how do we address the evolution of language with new slang and phrases continually emerging? These questions highlight the complexity of sexism detection. The cost and technical skills required to create a system that incorporates LLMs that can identify sexism make it unattainable for most individuals. We aim to simplify the process using pre-trained LLMs and prompts to address the EXIST lab tasks of classifying and labelling tweets.</p><p>In addition to the text classification in Tasks 1-3, we address the problem of identifying sexism in multi-modal formats for Task 4. Memes -ideas, images, or videos that are spread very quickly on the internet <ref type="bibr" target="#b15">[16]</ref> -exist not only in text form but also include any accompanying images. Therefore, combining text and the attached image (i.e., making the input multi-modal) can be more conducive to identifying whether a meme is sexist. Multi-modal models are usually proposed to deal with multi-modal datasets for classification tasks. Among existing multi-modal systems, Contrastive Language-Image Pre-Training (CLIP) <ref type="bibr" target="#b1">[2]</ref> is a powerful vision-and-language (VL) pre-trained model that can directly learn raw text about images. In addition, CLIP has the ability to map data of different modalities, text and images into a shared embedding space. Hence, CLIP has been shown to be a powerful tool for zero-shot image and text classification <ref type="bibr" target="#b1">[2]</ref>. Furthermore, CLIP can be beneficial for image-text feature fusion, which can boost model performance on natural language processing (NLP) downstream tasks such as text classification <ref type="bibr" target="#b16">[17]</ref> and multi-modal sarcasm detection <ref type="bibr" target="#b17">[18]</ref>. Motivated by the success of CLIP on various VL downstream tasks, this study aims to investigate the following research questions for Task 4:</p><p>• How effective is CLIP for zero-shot sexism identification?</p><p>• How can the naturally inherited multi-modal knowledge from pre-trained CLIP be extracted to identify sexism effectively?</p><p>Addressing the first research question, we proposed Prompt-CLIP for zero-shot sexism identification. For the latter question, we employed CLIP to perform supervised sexism classification.</p><p>Inspired by the impressive performance of multi-view CLIP for sarcasm detection in a previous study <ref type="bibr" target="#b17">[18]</ref>, we adopted multi-view CLIP for supervised sexism classification, namely, text-image multi-view CLIP (TIMV-CLIP) and proposed text-image multi-modal models via CLIP-Guided Learning (TI-CLIP) as a baseline. The paper is organized as follows. Details about the tasks participated in are described in Section 2. Section 3 provides details about the proposed approaches. In Section 4, we provide and discuss the results. Finally, we conclude in Section 5.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Tasks Addressed</head><p>The sEXism Identification in Social neTworks (EXIST) <ref type="bibr" target="#b18">[19]</ref> lab at the Conference and Labs of the Evaluation Forum (CLEF) 2024 <ref type="bibr" target="#b19">[20]</ref> aims to identify and characterize sexism using the learning with disagreements paradigm <ref type="bibr" target="#b20">[21,</ref><ref type="bibr" target="#b21">22,</ref><ref type="bibr" target="#b22">23]</ref>. This edition of the EXIST lab consists of sexism characterization on microblog posts (tweets) and memes.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Tasks 1-3: Sexism Characterization of Microblog Posts</head><p>• Task 1: Addresses sexism identification in tweets as a binary classification, requiring the system to classify whether a tweet is sexist (YES) or not (NO).</p><p>• Task 2: Focuses on determining the source intention in tweets as a multi-class classification, requiring the system to classify the tweet's intention as Direct, Reported, or Judgemental.</p><p>• Task 3: Involves sexism categorization in tweets as a multi-label classification, requiring the system to classify tweets into categories such as Ideological Inequality, Stereotyping Dominance, Objectification, Sexual Violence, and Misogyny-Non-Sexual Violence.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Task 4: Sexism Identification of Memes</head><p>While the above tasks address sexism identification in text, Task 4 deals with multi-modal input. Task 4 aims to address sexism identification as a binary classification, requiring the systems to classify whether a given meme is sexist or not. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Evaluation approaches</head><p>• Soft-Soft Evaluation: For systems that produce probabilities for each category, soft-soft evaluation is provided to compare the probabilities assigned by the systems with those assigned by the set of human annotators. The official evaluation metric is ICM-soft <ref type="bibr" target="#b23">[24,</ref><ref type="bibr" target="#b22">23]</ref>.</p><p>Additionally, Cross Entropy is also reported.</p><p>• Hard-Hard Evaluation: Hard labels are derived from the different annotators' labels through a probabilistic threshold computed for each task. Hard-hard evaluation is provided to evaluate systems that return Hard labels as output by comparing against a ground truth that combines multiple annotations into one. The original ICM <ref type="bibr" target="#b24">[25]</ref> and 𝐹 1 score are used as evaluation metrics.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Proposed Approaches</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Unsupervised In-Context Learning for Sexism Characterization in Microblog Posts</head><p>Our goal was to examine the procedure of developing a functional solution with readily available LLMs while minimizing the manual effort required from the practitioner. As shown in Figure <ref type="figure">1</ref>, the basic architecture involves giving the researcher a set of labeling or classification tasks and asking the LLM to generate an accurate output. To ensure that the responses followed the predefined criteria for each task, the outputs were systematically stored in a JSON format and manually inspected for errors in the "value" field. Responses such as "YES", "YES", or variations with additional text or punctuation like "Yes, the ... is sexist" required manual corrections to conform to the expected format. Incidences of token limit rates that resulted in "HTTP" errors were addressed by re-running the task for the affected tweet using its unique ID. These occurrences were uncommon, making manual correction a more efficient solution than an automated task given the time constraint. The prompts used for runs submitted to Tasks 1, 2, and 3 were designed with multiple parts:</p><p>• Definition of the underlying concept being addressed in the task (e.g., sexism): Sexism, prejudice or discrimination based on sex or gender, especially against women and girls. Although its origin is unclear, the term sexism emerged from the "second-wave" feminism of the 1960s through '80s and was most likely modeled on the civil rights movement's term racism (prejudice or discrimination based on race). Sexism can be a belief that one sex is superior to or more valuable than another sex. It imposes limits on what men and boys can and should do and what women and girls can and should do. The concept of sexism was originally formulated to raise consciousness about the oppression of girls and women, although by the early 21st century it had sometimes been expanded to include the oppression of any sex, including men and boys, intersex people, and transgender people.</p><p>• Instruction to Address Task and to Obtain Consistent Outputs: You are a robot who detects sexism from text given in the prompt.</p><p>• Perspectivism:</p><p>- • Instance to Classify:</p><formula xml:id="formula_0">#### [tweet] ####</formula><p>An example of a prompt submitted and output obtained from gpt-4-turbo to classify instance with id_EXIST 600090 using RMIT-IR_3 for Task 2 (Soft):</p><p>• Input: Sexism, prejudice or discrimination based on sex or gender, especially against women and girls. Although its origin is unclear, the term sexism emerged from the "second-wave" feminism of the 1960s through '80s and was most likely modeled on the civil rights movement's term racism (prejudice or discrimination based on race). Sexism can be a belief that one sex is superior to or more valuable than another sex. To find the distribution of responses, we initially tried to use GPT to figure out the likelihood percentage. Unfortunately, GPT only gave absolute values (either 100 or 0) or a consistent split of 70/30 most of the time. We directed the model to generate six responses for each tweet, which matched the number of annotators per tweet. For example, a set of responses like "YES", "NO", "YES", "NO", "YES", and "NO" would result in a calculated distribution of 50%. Additionally, in our final submissions, experimental runs two and three included prompts that provided additional context, such as the annotators' gender or educational backgrounds. This was done to see if providing relevant background information would improve the LLM's ability to predict annotator responses. The formats for each task can be seen above along with an example prompt used in RMIT-IR_3 for Task 2 (Soft).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.1.">Runs Submitted to Tasks 1-3</head><p>We used OpenAI's API to submit prompts to the pre-trained model gpt-4-turbo-2024-04-09 <ref type="bibr" target="#b25">[26]</ref>. For each tweet in the test set, we instantiated the prompt from above by appending the textual content of the instance. We used the syntax #### [tweet] #### to provide explicit delimiters to the model. For Soft tasks, we asked for six instances and then created a distribution based on the frequency of the predicted labels. We experimented with multiple versions of prompt templates using the development set supplied by the EXIST organizers (we did not use the training set). We found that the following</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 1</head><p>Summary of the runs submitted to Tasks 1-3. The Output Format was according to the type of task (Soft and Hard) as detailed previously. The instance to classify was appended at the end of the prompt.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Run</head><p>Definition Instruction Perspectivism Output Format</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>RMIT-IR_1 -RMIT-IR_2</head><p>Level of Education RMIT-IR_3</p><p>Level of Education + Gender elements were especially effective in directing the model to concentrate on the specific task and to ensure the responses were properly formatted (i.e., single-word answers and capitalized):</p><p>• Employing a role-playing technique of framing the task with the prompt "You are a robot who detects sexism from text given in the prompt. "</p><p>• Giving explicit formatting instructions such as "Give me 6 answers with NO or YES. Format:</p><p>[NO], [YES]".</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Multi-modal Contrastive Learning for Sexism Identification on Memes</head><p>Inspired by the successful applications of CLIP <ref type="bibr" target="#b1">[2]</ref> for NLP <ref type="bibr" target="#b16">[17]</ref> and computer vision tasks <ref type="bibr" target="#b26">[27]</ref>, <ref type="bibr" target="#b27">[28]</ref>, we adopted CLIP for the sexism identification task (Task 4). Unlike conventional methods that rely heavily on labelled image-text pairs, CLIP is a cross-modality model pre-trained with 400M noisy image-text pairs collected from the internet to learn high-level semantic features. CLIP consists of two encoders that embed texts and images into a uniform mathematical space.</p><p>Then, for the matched image-text pair, CLIP is encouraged to maximize the cosine similarity between the embedding of the two modalities. Otherwise, the similarity is minimized for the model to find the most suitable paired images and texts. Our motivation for using CLIP-based learning for sexism identification is to capture cross-modal ambiguity by explicitly measuring the correlation between texts and images of targeted memes and to guide the feature-fusing and decision-making stages. We propose two supervised contrastive learning models based on CLIP: Text-Image multimodal model via CLIP-guided learning (TI-CLIP) and Text-Image Multi-View multi-modal model via CLIP-guided learning (TIMV-CLIP). The architecture of TIMV-CLIP is shown in Figure <ref type="figure" target="#fig_1">2</ref>. We also propose Prompt-CLIP to address zero-shot sexism classification and CLIP-based models for supervised sexism classification.</p><p>• TI-CLIP: The overall architecture of TI-CLIP consists of two feature encoding models used to encode texts and images. These embeddings are then combined into a multi-modal embedding before passing into a feedforward network for sexism classification.</p><p>• TIMV-CLIP: We adopted a novel multi-view CLIP framework (MV-CLIP) <ref type="bibr" target="#b17">[18]</ref> for sexism identification, namely TIMV-CLIP (Figure <ref type="figure" target="#fig_1">2</ref>). In addition to encoding image and text as TI-CLIP, TIMV-CLIP further considers modelling relationships across text and image  modality using a transformer encoder, which aims to capture the interaction across different modalities. Unlike MV-CLIP in a previous study <ref type="bibr" target="#b17">[18]</ref>, TIMV-CLIP employs BERT Base Multilingual (mBERT) <ref type="bibr" target="#b28">[29]</ref> to encode texts.</p><p>• Prompt-CLIP: Prompt-CLIP performs zero-shot sexism identification. Prompt-CLIP uses a pre-trained CLIP model to create a custom classifier without training and considers images as inputs. It further encodes pre-defined classes (sexism and not sexism) with more description, known as prompts, into a learned latent space, and compares their similarity to the image latent space. In this study, we used "an image contains no information about sexism" and "an image contains information about sexism and against women" as prompts for Prompt-CLIP. The pre-trained text encoder transforms the class names (e.g., prompts) into a text embedding vector, while the pre-trained Image Encoder embeds the image.</p><p>• Model Training: We first randomly split the training subset into training (80%) and validation (20%) for cross-validation purposes. We implemented TI-CLIP and TIMV-CLIP based on the Hugging Face library <ref type="bibr" target="#b29">[30]</ref> and adopted clip-vit-base-patch32 as the backbone. Both TI-CLIP and TIMV-CLIP were trained directly with Soft labels. We use Adam as an optimizer to optimize the parameters in both TI-CLIP and TIMV-CLIP models. After several trials with other hyperparameters, we selected the parameters that performed best on the validation set. Specifically, the batch size is 32. The learning rate for CLIP is 1e-6 and for the other parts is 5e-4. Finally, we use the dropout percentage of 0.3 and train the models for 10 epochs. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.1.">Runs Submitted to Task 4</head><p>The proposed multi-modal sexism identification models mainly focus on Soft label predictions.</p><p>For the Hard submissions, hard labels are directly assigned by applying the max function, i.e., based on the highest probability score. Table <ref type="table" target="#tab_3">2</ref> presents the submitted runs for Task 4, which can be summarized as follows:</p><p>• RMIT-IR_1: For the first submission, the trained TI-CLIP model was used to predict whether given memes are sexist or not sexist.</p><p>• RMIT-IR_2: We used the trained TIMV-CLIP to generate the second submission.</p><p>• RMIT-IR_3: Prompt-CLIP was used to predict Soft and Hard labels for the third submission.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Results and Discussion</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Tasks 1-3</head><p>The performance of our proposed approaches for Tasks 1, 2, and 3 are presented in Tables <ref type="table" target="#tab_5">3,  4</ref>, and 5, respectively. As shown in Table <ref type="table" target="#tab_4">3</ref>, simplifying the architecture required for creating a LLM that can identify sexism shows promise, as evidenced by the classification of English and Spanish tweets. Although the current model achieved a 49% ICM-Soft score, which is 17% lower than the best-performing run, this result indicates the potential to use prompting for classification tasks. The cost of this process, which included testing with a development dataset and submitting with a gold dataset, was close to $150 AUD. In particular, it did not require knowledge of cloud computing, expensive hardware, or much energy. The average time taken to produce an output was about 90 minutes. Looking at Table <ref type="table" target="#tab_4">3</ref>, Task 1, the use of the Task-specific Prompt yielded an ICM-Soft Norm score of 49%, securing the 23rd position overall. It is interesting to note that the inclusion of additional clues such as the annotator's education level and gender did not bolster performance; instead, it diminished the score. In particular, when looking at the Spanish test instances, the model scored higher across all Spanish test instances compared to all English test instances, despite being trained on an English-based GPT. This underscores the robust cross-lingual applicability of the model, showcasing its proficient handling of Spanish data despite its primary training on English. In Task 2, our analysis revealed challenges in multi-class classification as shown in Table <ref type="table" target="#tab_5">4</ref>. The approach yielded a 13% ICM-Soft Norm score, indicating considerable difficulty in discerning the intention of the tweets. The introduction of additional clues, such as the annotator's education level, generally led to a decline in performance. However, adding gender information resulted in a slight improvement, elevating the score from almost 0% to 3%. The results indicated that the approach performed more effectively in English without additional clues; however, its performance diminished once clues were introduced. Conversely, our analysis demonstrated that the GPT model exhibited greater efficacy with clues in Spanish, suggesting potential advantages in providing contextual information in non-English scenarios. Task 3 was also challenging with multi-label classification. The initial ICM-Soft Norm score, as shown in Table <ref type="table" target="#tab_6">5</ref>, stood at 11%. Following a similar trend to Task 2, the integration of clues such as the level of education of the annotator, resulted in a reduction in performance to 8%. Subsequently, when both education and gender clues were included, performance decreased further to 4%. The English scores are quite similar to Task 2. Notably from the initial performance on the Spanish dataset exceeded that of Task 2, but declined with the addition of educational clues and further declined with the incorporation of both education and gender clues. This observation underscores a consistent pattern of diminishing returns with the incorporation of more specific annotator information.</p><p>Our proposed approaches for the second and third runs involve implementing few-shot and in-context learning. The experiments for these runs were conducted using gpt-4-turbo, and future tests should include gpt-4o along with other pre-trained LLMs to determine their efficacy in this context. Our experiment results show that involving few-shot and in-context learning does not improve model performance on sexism identification in tweets (as shown in Tables <ref type="table" target="#tab_6">3-5</ref>). Although prompting requires less coding and understanding of LLMs, producing the exact desired response 100% of the time was challenging. The prompts had to be carefully designed to ensure that the GPT provided a single and consistent answer, especially when dealing with distribution. Although pre-processing text for an LLM is more complex than pre-processing answers from GPTs, ensuring the response is in the correct format is simpler. Another unexpected aspect of this architecture was the ability to assist the GPT with hints. We tested how adding biases by including the annotator's education level and gender affected its ability to classify or label tweets. This gave insights into how such biases can influence model performance and classification accuracy.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Task 4</head><p>Table <ref type="table" target="#tab_7">6</ref> present the results of the proposed approaches for Task 4. Among the proposed approaches, TIMV-CLIP performs best in all cases (English+Spanish, English, or Spanish test instances) considering both Soft and Hard evaluation scenarios. This indicates the importance of effectively utilizing deep interactions between texts and images of memes with CLIP. Furthermore, TIMV-CLIP achieved the best-performance model (RMIT-IR_2) on English test instances with an ICM-Soft Norm score of 0.4998, ranked first in the leaderboard considering the Soft evaluation (English test instances). This observation confirms the advantages of CLIP for text-image pair classification tasks. However, the performance of TIMV-CLIP has dropped in Spanish test instances, which leads to lower performance in all test instances (Spanish test instances). We believe using a translation component for Spanish text in memes could lead to better overall performance.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusions</head><p>This paper proposed unsupervised in-context learning with off-the-shelf pre-trained LLMs to address sexism characterization on microblog posts (Tasks 1, 2, and 3). Dealing with multi-modal inputs, we proposed multi-modal contrastive learning, including Prompt-CLIP, TI-CLIP, and TIMV-CLIP for sexism identification in memes (Task 4).</p><p>The results of our experiment demonstrated the effectiveness of TIMV-CLIP under the Learning with Disagreements regime, indicating the need to consider capturing sexism cues from different perspectives, including image, text, and image-text interactions.</p><p>Future work includes further experimentation with unsupervised In-Context Learning in other tasks or meta-tasks such as MonsterCLEF <ref type="bibr" target="#b30">[31]</ref>, and the inclusion of machine translation for multi-modal contrastive learning.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head></head><label></label><figDesc>Level of Education: For each response, consider the perspective of individuals representing the following study levels: [study_levels_annotators] -Level of Education and Gender: For each response, consider the perspective of individuals representing the following study levels: [study_levels_annotators] and gender: [gender_annotators]. • Output Format: -Task 1 (Soft): Give me 6 answers with NO or YES. Format: [NO], [YES] -Task 1 (Hard): Give me 1 answer with [NO] or [YES] -Task 2 (Soft): Give me 6 answers with NO, DIRECT, REPORTED or JUDGEMEN-TAL using commas for each answer. Example: [NO], [DIRECT], [REPORTED], [JUDGEMENTAL], [JUDGEMENTAL], [NO] -Task 2 (Hard): Give me 1 answer with NO, DIRECT, REPORTED or JUDGEMEN-TAL using commas for each answer. Example: [NO], [DIRECT], [REPORTED], [JUDGEMENTAL], [JUDGEMENTAL], [NO] -Task 3 (Soft): Give me 6 answers with NO, IDEOLOGICAL-INEQUALITY, STEREOTYPING-DOMINANCE, OBJECTIFICATION, SEXUAL-VIOLENCE, or MISOGYNY-NON-SEXUAL-VIOLENCE using commas for each answer. Example: [NO], [IDEOLOGICAL-INEQUALITY], [STEREOTYPING-DOMINANCE], [OBJEC-TIFICATION], [SEXUAL-VIOLENCE], [MISOGYNY-NON-SEXUAL-VIOLENCE] -Task 3 (Hard): Give me 1 answers with NO, IDEOLOGICAL-INEQUALITY, STEREOTYPING-DOMINANCE, OBJECTIFICATION, SEXUAL-VIOLENCE, or MISOGYNY-NON-SEXUAL-VIOLENCE using commas for each answer. Example: [NO], [IDEOLOGICAL-INEQUALITY], [STEREOTYPING-DOMINANCE], [OBJEC-TIFICATION], [SEXUAL-VIOLENCE], [MISOGYNY-NON-SEXUAL-VIOLENCE]</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: The architecture of our proposed multi-modal model TIMV-CLIP for supervised sexism identification on memes (Task 4).</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head></head><label></label><figDesc>It imposes limits on what men and boys can and should do and what women and girls can and shoulddo. The concept of sexism was originally formulated to raise consciousness about the oppression of girls and women, although by the early 21st century it had sometimes been expanded to include the oppression of any sex, including men and boys, intersex people, and transgender people. You are a robot who detects sexism from text given in the prompt. For each response, consider the perspective of individuals representing the following</figDesc><table /><note>study levels: ["High school degree or equivalent", "Bachelor's degree", "Bachelor's degree", "Bachelor's degree", "Bachelor's degree", "High school degree or equivalent"]. Give me 6 answers with NO, DIRECT, REPORTED or JUDGEMENTAL using commas for each answer. Example: [NO], [DIRECT], [REPORTED], [JUDGEMENTAL], [JUDGEMENTAL], [NO]. #### Girls, don't let anyone ever tell you, you're not as good as a man #gender #girlpower #equity #### • Output: [NO],[NO],[NO],[NO],[NO],[NO]</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 2</head><label>2</label><figDesc>Runs submitted to Task 4: sexism identification on memes.</figDesc><table><row><cell>Run</cell><cell>Model</cell></row><row><cell cols="2">RMIT-IR_1 TI-CLIP (feedforward network)</cell></row><row><cell cols="2">RMIT-IR_2 TIMV-CLIP (Transformer encoder)</cell></row><row><cell cols="2">RMIT-IR_3 Prompt-CLIP (zero-shot)</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 3</head><label>3</label><figDesc>Results of the proposed approaches for Task 1 (Soft).</figDesc><table><row><cell>Rank ICM-Soft ICM-Soft Norm Cross Entropy</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 4</head><label>4</label><figDesc>Results of the proposed approaches for Task 2 (Soft).</figDesc><table><row><cell>Rank ICM-Soft ICM-Soft Norm Cross Entropy</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 5</head><label>5</label><figDesc>Results of the proposed approaches for Task 3 (Soft).</figDesc><table><row><cell>Rank ICM-Soft ICM-Soft Norm</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_7"><head>Table 6</head><label>6</label><figDesc>Results of the proposed approaches for Task 4 (Soft and Hard).</figDesc><table><row><cell></cell><cell cols="6">Rank (Soft) ICM-Soft ICM-Soft Norm Cross Entropy ICM-Hard ICM-Hard Norm</cell><cell>F1 YES</cell></row><row><cell>All Test Instances (English + Spanish)</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>EXIST2024_gold</cell><cell>0</cell><cell>3.1107</cell><cell>1.0000</cell><cell>0.5852</cell><cell>0.9832</cell><cell>1.0000</cell><cell>1.0000</cell></row><row><cell>EXIST2024_majority_class</cell><cell>36</cell><cell>−2.3568</cell><cell>0.1212</cell><cell>4.4015</cell><cell>−0.4038</cell><cell>0.2947</cell><cell>0.6821</cell></row><row><cell>EXIST2024_minority_class</cell><cell>38</cell><cell>−3.5089</cell><cell>0.0000</cell><cell>5.5672</cell><cell>−0.6468</cell><cell>0.1711</cell><cell>0.0000</cell></row><row><cell>TI-CLIP (RMIT-IR_1)</cell><cell>29</cell><cell>−1.2819</cell><cell>0.2940</cell><cell>1.0128</cell><cell>−0.6468</cell><cell>0.1711</cell><cell>0.0000</cell></row><row><cell>TIMV-CLIP (RMIT-IR_2)</cell><cell>8</cell><cell>−0.3780</cell><cell>0.4392</cell><cell>0.9852</cell><cell>−0.0123</cell><cell>0.4938</cell><cell>0.6726</cell></row><row><cell>Prompt-CLIP (RMIT-IR_3)</cell><cell>24</cell><cell>−1.0894</cell><cell>0.3249</cell><cell>1.1206</cell><cell>−0.2601</cell><cell>0.3677</cell><cell>0.6040</cell></row><row><cell>English Test Instances</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>EXIST2024_gold</cell><cell>0</cell><cell>3.0794</cell><cell>1.0000</cell><cell>0.5528</cell><cell>0.9848</cell><cell>1.0000</cell><cell>1.0000</cell></row><row><cell>EXIST2024_majority_class</cell><cell>34</cell><cell>−2.2236</cell><cell>0.1390</cell><cell>4.4798</cell><cell>−0.4076</cell><cell>0.2931</cell><cell>0.6880</cell></row><row><cell>EXIST2024_minority_class</cell><cell>36</cell><cell>−3.1235</cell><cell>0.0000</cell><cell>5.4888</cell><cell>−0.6381</cell><cell>0.1761</cell><cell>0.0000</cell></row><row><cell>TI-CLIP (RMIT-IR_1)</cell><cell>33</cell><cell>−1.2889</cell><cell>0.2907</cell><cell>1.0115</cell><cell>−0.6381</cell><cell>0.1761</cell><cell>0.0000</cell></row><row><cell>TIMV-CLIP (RMIT-IR_2)</cell><cell>1</cell><cell>−0.0011</cell><cell>0.4998</cell><cell>0.9243</cell><cell>0.1536</cell><cell>0.5780</cell><cell>0.7250</cell></row><row><cell>Prompt-CLIP (RMIT-IR_3)</cell><cell>25</cell><cell>−1.0106</cell><cell>0.3359</cell><cell>1.1316</cell><cell>−0.2089</cell><cell>0.3940</cell><cell>0.5641</cell></row><row><cell>Spanish Test Instances</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>EXIST2024_gold</cell><cell>0</cell><cell>3.1360</cell><cell>1.0000</cell><cell>0.6160</cell><cell>0.9815</cell><cell>1.0000</cell><cell>1.0000</cell></row><row><cell>EXIST2024_majority_class</cell><cell>36</cell><cell>−2.4997</cell><cell>0.1014</cell><cell>4.3270</cell><cell>−0.4001</cell><cell>0.2962</cell><cell>0.6765</cell></row><row><cell>EXIST2024_minority_class</cell><cell>38</cell><cell>−3.9408</cell><cell>0.0000</cell><cell>5.6416</cell><cell>−0.6557</cell><cell>0.1660</cell><cell>0.0000</cell></row><row><cell>TI-CLIP (RMIT-IR_1)</cell><cell>29</cell><cell>−1.2730</cell><cell>0.2970</cell><cell>1.0141</cell><cell>−0.6557</cell><cell>0.1660</cell><cell>0.0000</cell></row><row><cell>TIMV-CLIP (RMIT-IR_2)</cell><cell>17</cell><cell>−0.7851</cell><cell>0.3748</cell><cell>1.0431</cell><cell>−0.1762</cell><cell>0.4103</cell><cell>0.6192</cell></row><row><cell>Prompt-CLIP (RMIT-IR_3)</cell><cell>27</cell><cell>−1.1903</cell><cell>0.3102</cell><cell>1.1101</cell><cell>−0.3140</cell><cell>0.3400</cell><cell>0.6332</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">We acknowledge that the classification of sex and gender into two categories is a simplification of people's identities.</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This research has been carried out in the unceded lands of the Woi Wurrung and Boon Wurrung peoples of the eastern Kulin Nation. We pay our respects to their Ancestors and Elders, past, present, and emerging. This research is partially supported by the Australian Research Council (ARC, project nr. DE200100064 and CE200100005).</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<author>
			<persName><forename type="first">Q</forename><surname>Dong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Zheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Sui</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2301.00234</idno>
		<title level="m">A Survey on In-context Learning</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Learning Transferable Visual Models From Natural Language Supervision</title>
		<author>
			<persName><forename type="first">A</forename><surname>Radford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">W</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Hallacy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ramesh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Goh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sastry</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Askell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Mishkin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Clark</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Krueger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sutskever</surname></persName>
		</author>
		<ptr target="http://proceedings.mlr.press/v139/radford21a.html" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 38th International Conference on Machine Learning, ICML 2021</title>
				<editor>
			<persName><forename type="first">M</forename><surname>Meila</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Zhang</surname></persName>
		</editor>
		<meeting>the 38th International Conference on Machine Learning, ICML 2021<address><addrLine>PMLR</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021-07">July 2021. 2021</date>
			<biblScope unit="volume">139</biblScope>
			<biblScope unit="page" from="8748" to="8763" />
		</imprint>
	</monogr>
	<note>Proceedings of Machine Learning Research</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Social Media and Mental Health</title>
		<author>
			<persName><forename type="first">L</forename><surname>Braghieri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Levy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Makarin</surname></persName>
		</author>
		<idno type="DOI">10.1257/aer.20211218</idno>
	</analytic>
	<monogr>
		<title level="j">American Economic Review</title>
		<imprint>
			<biblScope unit="volume">112</biblScope>
			<biblScope unit="page" from="3660" to="3693" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Young Adults&apos; Folk Theories of How Social Media Harms Its Users</title>
		<author>
			<persName><forename type="first">R</forename><surname>Young</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Kananovich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">G</forename><surname>Johnson</surname></persName>
		</author>
		<idno type="DOI">10.1080/15205436.2021.1970186</idno>
	</analytic>
	<monogr>
		<title level="j">Mass Communication and Society</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<biblScope unit="page" from="23" to="46" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">The Problem of Anti-feminist &apos;Manfluencer&apos; Andrew Tate in Australian Schools: Women Teachers&apos; Experiences of Resurgent Male Supremacy</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">R</forename><surname>Stephanie Wescott</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhao</surname></persName>
		</author>
		<idno type="DOI">10.1080/09540253.2023.2292622</idno>
	</analytic>
	<monogr>
		<title level="j">Gender and Education</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<biblScope unit="page" from="167" to="182" />
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<ptr target="https://dictionary.cambridge.org/dictionary/english/sexism" />
		<title level="m">Cambridge Dictionary, Definition of sexism</title>
				<imprint>
			<date type="published" when="2024-07-04">2024. 2024-07-04</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">The State and the Oppression of Women</title>
		<author>
			<persName><forename type="first">M</forename><surname>Mcintosh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Feminism and Materialism (RLE Feminist Theory)</title>
				<imprint>
			<publisher>Routledge</publisher>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="254" to="289" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<author>
			<persName><forename type="first">G</forename><surname>Masequesmay</surname></persName>
		</author>
		<ptr target="https://www.britannica.com/topic/sexism" />
		<title level="m">Sexism | Definition, Types, Examples, &amp; Facts | Britannica</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Ambivalent sexism</title>
		<author>
			<persName><forename type="first">P</forename><surname>Glick</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">T</forename><surname>Fiske</surname></persName>
		</author>
		<idno type="DOI">10.1016/S0065-2601(01)80005-8</idno>
	</analytic>
	<monogr>
		<title level="j">Advances in Experimental Social Psychology</title>
		<imprint>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="115" to="188" />
			<date type="published" when="2001">2001</date>
			<publisher>Academic Press</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Misogynistic Tweets Correlate with Violence against Women</title>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">R</forename><surname>Blake</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>O'dean</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">F</forename><surname>Denson</surname></persName>
		</author>
		<idno type="DOI">10.1177/0956797620968529</idno>
	</analytic>
	<monogr>
		<title level="j">Psychological science</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<biblScope unit="page" from="315" to="325" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Online Misogyny</title>
		<author>
			<persName><forename type="first">K</forename><surname>Barker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Jurasz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of International Affairs</title>
		<imprint>
			<biblScope unit="volume">72</biblScope>
			<biblScope unit="page" from="95" to="114" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Multi-class Hate Speech Detection in the Norwegian Language Using FAST-RNN and Multilingual Fine-Tuned Transformers</title>
		<author>
			<persName><forename type="first">E</forename><surname>Hashmi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">Y</forename><surname>Yayilgan</surname></persName>
		</author>
		<idno type="DOI">10.1007/s40747-024-01392-5</idno>
	</analytic>
	<monogr>
		<title level="j">Complex &amp; Intelligent Systems</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page" from="4535" to="4556" />
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">UMUTeam at EXIST 2023: Sexism Identification and Categorisation Fine-tuning Multilingual Large Language Models</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>García-Díaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Pan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Valencia-García</surname></persName>
		</author>
		<ptr target="https://ceur-ws.org/Vol-3497/paper-080.pdf" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2023 -Conference and Labs of the Evaluation Forum</title>
				<editor>
			<persName><forename type="first">M</forename><surname>Aliannejadi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Vlachos</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="985" to="999" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Multi-task Learning Neural Framework for Categorizing Sexism</title>
		<author>
			<persName><forename type="first">H</forename><surname>Abburi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Parikh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Chhaya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Varma</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.csl.2023.101535</idno>
	</analytic>
	<monogr>
		<title level="j">Comput. Speech Lang</title>
		<imprint>
			<biblScope unit="volume">83</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Overview of EXIST 2023: sEXism Identification in Social neTworks</title>
		<author>
			<persName><forename type="first">L</forename><surname>Plaza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Carrillo-De Albornoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Morante</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Amigó</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Spina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-031-28241-6_68</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of ECIR&apos;</title>
				<meeting>ECIR&apos;</meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="volume">23</biblScope>
			<biblScope unit="page" from="593" to="599" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<ptr target="https://dictionary.cambridge.org/dictionary/english/meme" />
		<title level="m">Cambridge Dictionary, Definition of meme</title>
				<imprint>
			<date type="published" when="2024-07-04">2024. 2024-07-04</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">CLIPText: A New Paradigm for Zero-shot Text Classification</title>
		<author>
			<persName><forename type="first">L</forename><surname>Qin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Che</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2023.findings-acl.69</idno>
	</analytic>
	<monogr>
		<title level="m">Findings of the Association for Computational Linguistics: ACL 2023, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">A</forename><surname>Rogers</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Boyd-Graber</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Okazaki</surname></persName>
		</editor>
		<meeting><address><addrLine>Toronto, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="1077" to="1088" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">MMSD2.0: Towards a Reliable Multi-modal Sarcasm Detection System</title>
		<author>
			<persName><forename type="first">L</forename><surname>Qin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Cai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Liang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Che</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Xu</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2023.findings-acl.689</idno>
	</analytic>
	<monogr>
		<title level="m">Findings of the Association for Computational Linguistics: ACL 2023, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">A</forename><surname>Rogers</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Boyd-Graber</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Okazaki</surname></persName>
		</editor>
		<meeting><address><addrLine>Toronto, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="10834" to="10845" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Plaza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Carrillo-De Albornoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Amigó</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Morante</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Spina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chulvi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Maeso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Ruiz</surname></persName>
		</author>
		<ptr target="http://nlp.uned.es/exist2024/" />
		<title level="m">EXIST: sEXism Identification in Social neTworks</title>
				<imprint>
			<date type="published" when="2024-07-04">2024. 2024-07-04</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<ptr target="https://www.clef-initiative.eu/" />
		<title level="m">Conference and Labs of the Evaluation Forum (CLEF)</title>
				<imprint>
			<date type="published" when="2024-07-04">2024. 2024-07-04</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">EXIST 2024: sEXism Identification in Social neTworks and Memes</title>
		<author>
			<persName><forename type="first">L</forename><surname>Plaza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Carrillo-De Albornoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Amigó</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Morante</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Spina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chulvi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Maeso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Ruiz</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-031-56069-9_68</idno>
	</analytic>
	<monogr>
		<title level="m">Advances in Information Retrieval: 46th European Conference on Information Retrieval, ECIR 2024</title>
				<meeting><address><addrLine>Glasgow, UK; Part V; Berlin, Heidelberg</addrLine></address></meeting>
		<imprint>
			<publisher>Springer-Verlag</publisher>
			<date type="published" when="2024">March 24-28, 2024. 2024</date>
			<biblScope unit="page" from="498" to="504" />
		</imprint>
	</monogr>
	<note>Proceedings</note>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Overview of EXIST 2024 -Learning with Disagreement for Sexism Identification and Characterization in Social Networks and Memes</title>
		<author>
			<persName><forename type="first">L</forename><surname>Plaza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Carrillo-De-Albornoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Ruiz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Maeso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chulvi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Amigó</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Morante</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Spina</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fifteenth International Conference of the CLEF Association</title>
				<meeting><address><addrLine>CLEF</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2024">2024. 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Overview of EXIST 2024 -Learning with Disagreement for Sexism Identification and Characterization in Social Networks and Memes (Extended Overview)</title>
		<author>
			<persName><forename type="first">L</forename><surname>Plaza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Carrillo-De-Albornoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Ruiz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Maeso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chulvi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Amigó</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Morante</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Spina</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2024 -Conference and Labs of the Evaluation Forum</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">G S</forename><surname>Herrera</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Overview of EXIST 2023 -Learning with Disagreement for Sexism Identification and Characterization (Extended Overview)</title>
		<author>
			<persName><forename type="first">L</forename><surname>Plaza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Carrillo-De-Albornoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Morante</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Amigó</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Spina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<ptr target="https://ceur-ws.org/Vol-3497/paper-070.pdf" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2023 -Conference and Labs of the Evaluation Forum</title>
				<editor>
			<persName><forename type="first">M</forename><surname>Aliannejadi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Vlachos</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="813" to="854" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Evaluating Extreme Hierarchical Multi-label Classification</title>
		<author>
			<persName><forename type="first">E</forename><surname>Amigó</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Delgado</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2022.acl-long.399</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">S</forename><surname>Muresan</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Villavicencio</surname></persName>
		</editor>
		<meeting>the 60th Annual Meeting of the Association for Computational Linguistics<address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="5809" to="5819" />
		</imprint>
	</monogr>
	<note>: Long Papers), Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b25">
	<monogr>
		<title/>
		<author>
			<persName><surname>Openai</surname></persName>
		</author>
		<ptr target="https://help.openai.com/en/articles/8555510-gpt-4-turbo-in-the-openai-api" />
		<imprint>
			<date type="published" when="2024-07-04">2024. 2024-07-04</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">CMA-CLIP: Cross-Modality Attention Clip for Text-Image Classification</title>
		<author>
			<persName><forename type="first">J</forename><surname>Fu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Xie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-C</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Wang</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICIP46576.2022.9897323</idno>
	</analytic>
	<monogr>
		<title level="m">IEEE International Conference on Image Processing (ICIP)</title>
				<imprint>
			<date type="published" when="2022">2022. 2022</date>
			<biblScope unit="page" from="2846" to="2850" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">CLIP-Art: Contrastive Pre-training for Fine-Grained Art Classification</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">V</forename><surname>Conde</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Turgutlu</surname></persName>
		</author>
		<idno type="DOI">10.1109/CVPRW53098.2021.00444</idno>
	</analytic>
	<monogr>
		<title level="m">IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)</title>
				<imprint>
			<date type="published" when="2021">2021. 2021</date>
			<biblScope unit="page" from="3951" to="3955" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding</title>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-W</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/N19-1423</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</title>
		<title level="s">Long and Short Papers</title>
		<editor>
			<persName><forename type="first">J</forename><surname>Burstein</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Doran</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Solorio</surname></persName>
		</editor>
		<meeting>the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies<address><addrLine>Minneapolis, Minnesota</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="4171" to="4186" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">Transformers: State-of-the-Art Natural Language Processing</title>
		<author>
			<persName><forename type="first">T</forename><surname>Wolf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Debut</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Sanh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chaumond</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Delangue</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Moi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Cistac</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Rault</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Louf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Funtowicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Davison</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Shleifer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Von Platen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Jernite</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Plu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">Le</forename><surname>Scao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gugger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Drame</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Lhoest</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rush</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2020.emnlp-demos.6</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">Q</forename><surname>Liu</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Schlangen</surname></persName>
		</editor>
		<meeting>the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="38" to="45" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<monogr>
		<author>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Karlgren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Müller</surname></persName>
		</author>
		<ptr target="https://monsterclef.dei.unipd.it/" />
		<title level="m">MonsterCLEF: One Lab to Rule Them All</title>
				<imprint>
			<date type="published" when="2024-07-04">2024. 2024-07-04</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
