<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Small Language Models and Large Language Models in Oppositional Thinking Analysis: Capabilities, Biases and Challenges Notebook for PAN at CLEF 2024</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Álvaro</forename><surname>Huertas-García</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Department of Computer System Engineering</orgName>
								<orgName type="institution">Polytechnic University of Madrid</orgName>
								<address>
									<addrLine>Calle de Alan Turing</addrLine>
									<postCode>28031</postCode>
									<settlement>Madrid</settlement>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">Fundación Tecnológica Advantx -Funditec</orgName>
								<address>
									<addrLine>Paseo de la Castellana</addrLine>
									<postCode>28046</postCode>
									<settlement>Madrid</settlement>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Carlos</forename><surname>Martí-González</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">Fundación Tecnológica Advantx -Funditec</orgName>
								<address>
									<addrLine>Paseo de la Castellana</addrLine>
									<postCode>28046</postCode>
									<settlement>Madrid</settlement>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Javier</forename><surname>Muñoz</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">Fundación Tecnológica Advantx -Funditec</orgName>
								<address>
									<addrLine>Paseo de la Castellana</addrLine>
									<postCode>28046</postCode>
									<settlement>Madrid</settlement>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Enrique</forename><surname>De</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Miguel</forename><surname>Ambite</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">Fundación Tecnológica Advantx -Funditec</orgName>
								<address>
									<addrLine>Paseo de la Castellana</addrLine>
									<postCode>28046</postCode>
									<settlement>Madrid</settlement>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Small Language Models and Large Language Models in Oppositional Thinking Analysis: Capabilities, Biases and Challenges Notebook for PAN at CLEF 2024</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">84EA8AB9D718D1A2B8D83623AF98F669</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:59+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>PAN 2024</term>
					<term>Oppositional Thinking Analysis</term>
					<term>Transformers</term>
					<term>Mamba</term>
					<term>LLM</term>
					<term>Claude</term>
					<term>Bias</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The proliferation of misinformation and conspiracy theories needs robust methods to differentiate legitimate critical discourse from harmful conspiratorial narratives. This study investigates discerning critical messages from conspiracy theories within COVID-19 discussions on Telegram. Preserving information integrity on social media impacts vital public discourse on health, politics, and science.</p><p>The research employs two distinct approaches: linguistic style classification and contextual knowledge classification. The former leverages a diverse ensemble of Small Language Models (SLMs), Large Language Models (LLMs), and State-Space Models (SSMs), while the latter harnesses the capabilities of the Claude 2.0 Opus model for contextual analysis.</p><p>Empirical evaluations demonstrate that the SLM models using Matryoshka embedding and Mamba (SSM) models exhibit superior performance for the English language dataset, achieving a Matthews Correlation Coefficient (MCC) of 0.793. For the Spanish dataset, the Spanish BERT baseline (SLM) attains an MCC of 0.699. Notably, a multilingual model trained on a balanced combination of English and Spanish data outperforms its monolingual counterparts, with the multilingual-e5-large model (LLM) achieving an MCC of 0.768 for English and 0.725 for Spanish. This finding underscores the potential of multilingual models to mitigate the "curse of multilinguality, " where performance often degrades on low-resource languages. However, the suboptimal performance of the Claude 2.0 Opus model, exhibiting a tendency to classify texts as conspiracy-related, highlights inherent biases that require further investigation.</p><p>Overall, this study contributes to the development of advanced models that can effectively differentiate critical thinking from conspiratorial narratives in various linguistic contexts. Future research should prioritize identifying and addressing biases in large language models to ensure fair treatment of diverse perspectives, as well as to preserve freedom of expression and ensure fair representation of narratives.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The proliferation of misinformation and conspiracy theories has become a significant challenge in today's digital age, impacting vital aspects of public discourse such as health, politics, and scientific discourse <ref type="bibr" target="#b0">[1]</ref>. Conspiracy beliefs can shape human behaviour and decision-making processes, making understanding the cognitive styles and personality traits associated with such beliefs is crucial. Extensive psychological research has identified numerous predictors of conspiracy beliefs, including personality factors like low agreeableness and high openness to experience <ref type="bibr" target="#b0">[1]</ref>. Moreover, studies on cognitive styles have revealed a correlation between belief in conspiracy theories and lower analytic thinking coupled with higher intuitive thinking <ref type="bibr" target="#b1">[2]</ref>.</p><p>Understanding these human aspects of conspiracy beliefs is not merely an academic exercise; it has far-reaching implications. This knowledge can inform behavioural interventions to mitigate the spread of misinformation <ref type="bibr" target="#b2">[3]</ref>. Furthermore, integrating cognitive and personality factors into natural language processing (NLP) models can enhance their accuracy in distinguishing between critical and conspiratorial narratives, ultimately improving their performance and reliability.</p><p>In the realm of automatic content moderation, the challenge of distinguishing between conspiracy theories and critical thinking in NLP models has emerged as a vital area of study. The prevalence of conspiratorial content has escalated the need for robust methodologies that can accurately differentiate between legitimate critical discourse and harmful conspiracy narratives. Maintaining the integrity of information shared across social media platforms and other digital forums is crucial for preserving the credibility of public discourse.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Background</head><p>While the focus on this topic remains relatively limited, related studies provide valuable insights into methodologies and applications in adjacent areas. For instance, a significant contribution by <ref type="bibr" target="#b3">[4]</ref> presents a framework for detecting conspiracy theories on Twitter using a novel recurrent model called BORJIS, highlighting the efficacy and challenges of NLP techniques in identifying that conspiratorial content is often found within vast amounts of social media data. Similarly, <ref type="bibr" target="#b4">[5]</ref> explored fake news detection related to COVID-19 and 5G conspiracy theories using BERT embeddings and Graph Neural Networks, showcasing advanced NLP techniques for distinguishing misinformation from legitimate critical analysis.</p><p>Other studies follow a different approach focusing on tracking the spread across social networks, such as the FacTeR-Check semi-automated fact-checking tool that uses semantic similarity and natural language inference (NLI) to monitor the evolution of misinformation or disinformation on online social networks <ref type="bibr" target="#b2">[3]</ref>. Additionally, the use of camouflage for content evasion has also been reported, and works have developed multilingual NER NLP models to counter these strategies, like the "pyleetspeak" tool for simulating word camouflage and a NER Transformer model for its detection <ref type="bibr" target="#b5">[6]</ref>.</p><p>Furthermore, the research conducted by <ref type="bibr" target="#b6">[7]</ref> explores the potential of NLP techniques in fostering critical thinking skills within educational settings. It offers valuable insights into the systematic instruction and assessment of critical thinking, specifically in comparison to conspiratorial thinking. Finally, <ref type="bibr" target="#b7">[8]</ref> emphasize the importance of explicit theorization in developing models that can accurately differentiate between critical and conspiratorial thinking in their paper on gender bias in NLP research.</p><p>While significant progress has been made in the field, there is still much to explore in analyzing oppositional thinking using NLP. This research article addresses this issue in both English and Spanish, contributing to the development of more sophisticated NLP systems for real-world scenarios.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Competition Description</head><p>The competition, titled "Oppositional Thinking Analysis: Conspiracy vs Critical Narratives" is part of the PAN at CLEF 2024 event <ref type="bibr" target="#b8">[9,</ref><ref type="bibr" target="#b9">10]</ref>. Our focus is on the first subtask, which involves analyzing texts from the Telegram platform related to the COVID-19 pandemic. The objective is to perform a binary classification to differentiate between two types of narratives:</p><p>• Critical comment: Messages that question major decisions in the public health domain without promoting a conspiracist mentality. These are critical opinions based on information that may not be commonly accepted but do not imply secret plots or malevolent intentions. • Conspiracy comment: Messages that portray the pandemic or public health decisions as results of malevolent conspiracies by secret, influential groups. These messages often encourage distrust based on unverified or poorly explained evidence.</p><p>The official evaluation metric for this subtask is the Matthews Correlation Coefficient (MCC). MCC is a measure of the quality of binary classifications, providing a balanced evaluation even when the </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Methodology</head><p>Our methodology is based on two main approaches: one that classifies texts according to their linguistic style and content, which we refer to as the Linguistic Style Classification Approach; and another that uses the input text, combined with contextual knowledge and reasoning from large language models (LLMs), referred to as the Contextual Knowledge Classification Approach.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Linguistic Style Classification Approach</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.1.">Dataset Preprocessing</head><p>The dataset comprises texts in both English and Spanish, categorized into CRITICAL and CONSPIRACY narratives. The English dataset consists of 2,621 CRITICAL texts and 1,379 CONSPIRACY texts. The Spanish dataset includes 2,538 CRITICAL texts and 1,462 CONSPIRACY texts. Both datasets were divided into training (80%) and validation (20%) sets using a random seed of 42. The preprocessing involved analyzing the prevalence of URLs, emojis, and text length distributions. URLs were removed to standardize the text data. The text length distributions for the English dataset were found to be 743±740 characters for CONSPIRACY and 476±479 characters for CRITICAL. For the Spanish dataset, the distributions were 1112±946 characters for CONSPIRACY and 641±577 characters for CRITICAL. These distributions are illustrated in Figure <ref type="figure" target="#fig_0">1</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.2.">Models</head><p>We employed a diverse range of models, both monolingual and multilingual. Except for Mamba, all models are based on the Transformer architecture. The significance of Transformer models lies in their attention mechanism, which allows them to efficiently handle dependencies in long sequences and capture intricate patterns within the data. According to Vaswani et al. <ref type="bibr" target="#b10">[11]</ref>, the self-attention mechanism of Transformers enables them to dynamically weigh the importance of different tokens in a sequence, making them highly effective for various NLP tasks. Mamba, in contrast, is an advanced state-space model (SSM) designed for efficient handling of complex sequences with large datasets, as detailed by Gu and Dao <ref type="bibr" target="#b11">[12]</ref>.</p><p>Below, we list the models used in our research along with brief descriptions: Monolingual</p><p>• BERT-base-uncased <ref type="bibr" target="#b12">[13]</ref>: A foundational model that effectively applies Transformers at scale, expanding our understanding of linguistic context. We selected the largest variant to ensure a comprehensive analysis and to compare historical model design evolution. • DistilBERT<ref type="foot" target="#foot_0">1</ref> : A compact version of BERT by Hugging Face, offering a smaller and faster alternative while maintaining similar performance. Suitable for various NLP tasks. • Nomic: Nomic Embed <ref type="bibr" target="#b13">[14]</ref> innovates in embedding techniques to provide dynamic, contextaware representations, surpassing leading models as of February 2024. With a compact size, low memory usage, and advanced training methods, Nomic Embed efficiently processes up to 8192 tokens, making it ideal for analyzing extensive online materials. • DistilRoBERTa <ref type="bibr" target="#b14">[15]</ref>: A faster and smaller version of RoBERTa, trained on the same corpus in a self-supervised manner using BERT as a teacher. • twitter-roberta-base-sentiment-latest <ref type="bibr" target="#b15">[16]</ref>: A RoBERTa-base model fine-tuned for sentiment analysis using tweets from January 2018 to December 2021, benchmarked with TweetEval. • all-MiniLM-L6-v2<ref type="foot" target="#foot_1">2</ref> : A Transformer model trained with contrastive loss on 1B sentence pairs to encode sentences and short paragraphs into a dense vector space of 384 dimensions, suitable for tasks like clustering or semantic search. • mxbai-embed-large-v1 <ref type="bibr" target="#b16">[17]</ref>: A powerful English embedding model known for its efficient size and high performance. Using Matryoshka Embedding <ref type="bibr" target="#b17">[18]</ref>, it trains hidden layers to generate high-quality embeddings independently of higher layers, reducing both the number of layers and embedding dimensions. Ranked in the top 25 on the MTEB leaderboard <ref type="foot" target="#foot_2">3</ref> for sentence embedding tasks, it outperforms commercial models like OpenAI's text-embedding-3-large, making it a top choice for our research. • Mamba<ref type="foot" target="#foot_3">4</ref>  <ref type="bibr" target="#b11">[12]</ref>: An advanced state-space model designed for efficient handling of complex sequences with large datasets. It uses a selection mechanism to decide whether to propagate or discard information based on token relevance, providing a viable method for assessing intricate controversial and critical comments on social media. • dccuchile/bert-base-spanish-wwm-uncased <ref type="bibr" target="#b18">[19]</ref>: Also known as BETO, this is a BERT model trained on a large Spanish corpus using a vocabulary of about 31k BPE subwords constructed with SentencePiece.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Multilingual</head><p>• XLM-RoBERTa: A scaled cross-lingual multilingual sentence encoder version of the RoBERTa model, trained on 2.5TB of data across 100 languages filtered from Common Crawl. • LLAMA 2: A family of pre-trained and fine-tuned large language models (LLMs) by Meta AI, useful for various research and commercial purposes. • multilingual-e5: Developed at Microsoft, this sophisticated embedding model excels in tasks requiring robust text representation, such as information retrieval, semantic textual similarity, and text reranking. Initialized from xlm-roberta-large, it is continually trained on a mixture of multilingual datasets, supporting 100 languages from xlm-roberta with potential performance degradation for low-resource languages.</p><p>In this style-based strategy, all these models are employed as the encoder body of the texts to which a layer of 1024 classifier neurons is added. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.3.">Hyperparameter Tuning, Importance, and Correlation</head><p>For hyperparameter tuning, we used Bayesian optimization, which leverages prior evaluations to guide its search process, enhancing model performance. Table <ref type="table" target="#tab_0">1</ref> lists the explored hyperparameters, including their ranges and sampling distributions. We analyzed the importance and correlation of hyperparameters with the Matthews Correlation Coefficient (MCC). Correlation measures the linear relationship between hyperparameters and MCC, indicating how changes in hyperparameters affect performance.</p><p>Additionally, we calculated an importance metric exploiting the feature importance of a random forest model, based on the idea that the more important features appear more often in the trees of the forest. Hyperparameters served as input features, with MCC as the target output. This provided feature importance values, showing each hyperparameter's contribution to predicting performance. These analyses offer insights into how hyperparameters influence model performance.</p><p>The experimental tracking can be consulted in Weight and Biases<ref type="foot" target="#foot_4">5</ref> </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Contextual Knowledge Classification Approach</head><p>For this approach, we utilized the Claude 2.0 Opus model for zero-shot classification. This approach relies on prompt engineering using tempeareture equals to 1 and without fine-tuning the model, leveraging the extensive knowledge of language and context up to early 2023. Additionally, its multilingual capability is well-suited to this task, as approximately 10% of the data used was non-English, according to Anthropic. This model was accessed via Anthropic's public API before May 6. Below is the final prompt used for the Zero-shot classification:</p><p>Claude 2.0 Opus Prompt Your role is to analyze text inputs to identify whether they represent critical commentary or conspiracy theories, each with distinct characteristics: Critical Commentary: Definition: Critical messages that question major decisions in the public health domain, but do not promote a conspiracist mentality. It is an opinion, it may not be correct but do not consider that the revendication belongs to a secret or a plot against the population in terms of influential groups. It can be a critic based on information that may not be the common opinion, and it could be wrong, but it is expressing a point of view that another can criticize. Characteristics: Applicability: Applies even when the topic might be susceptible to conspiratorial interpretations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Conspiracy Commentary:</head><p>Definition: Messages that view the pandemic or public health decisions as a result of a malev-olent conspiracy by secret, influential groups. It can be an opinion but the main problem is that it tries to convince you to distrust based on evidences that are not well trusted or explained and leave open the door to be distrustful instead of being critical based on information that may not be the common opinion. Characteristics:</p><p>• Suspicion and Paranoia: Thrives on distrust of official narratives and institutions.</p><p>• Simplistic Explanations: Oversimplifies complexities by attributing them to the actions of a few. • Resistance to Evidence: Dismisses contrary evidence as part of the cover-up.</p><p>You are required to utilize web browser research extensively to verify claims and gather context before making your classification. Be sure to adhere strictly to the output format, especially in reporting URLs used in your research to ensure transparency and accountability. Additional Instruction:</p><p>• Always use a web browser to search for information related to the text. Your classification should be informed by credible online sources. Include URLs of these sources in your explanation to validate your findings and reasoning. • Maintain neutrality in your classification process. Do not classify a text as "CONSPIR-ACY" solely because the topic is related to commonly misunderstood or hot-button issues. Instead, use clear evidence from the text and supporting information from web searches to distinguish between critical perspectives and actual conspiracy theories. Include URLs of these sources in your explanation to validate your findings and reasoning.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Task Requirements:</head><p>• Classify the narrative of the text based on the categories above.</p><p>• Determine the main topics of the text in 2-3 words.</p><p>• Assign a Confidence Score from 0 to 1, indicating the certainty of your classification.</p><p>• Your explanation must reflect how the information sourced online influenced your classification and must include URLs for verification.</p><p>Output Format: (It is crucial that the output strictly follows this format) { "Prediction": "CATEGORY_NAME", "Confidence": [Confidence Score], "Topic": ["topic1", "topic2"], "Reason": "A concise explanation based on the characteristics with URLs of the sources used." } It is crucial that the output strictly follows this format.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Experiments and Results</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Training and Developing Results</head><p>This section presents the results obtained from developing and evaluating various models. As shown in Table <ref type="table" target="#tab_1">2</ref>, the best monolingual model for English are mxbai-embed-large-v1 and Mamba 370m model, achieving an MCC of 0.793, while the best model for Spanish is bert-base-spanish-wwm-uncased, with an MCC of 0.699. As shown in Table <ref type="table" target="#tab_1">2</ref>, the superior performance of the mxbai-embed-large-v1 model underscores its effectiveness in encoding texts into embeddings. This model employs the Matryoshka Embedding technique <ref type="bibr" target="#b17">[18]</ref>, where each layer is trained to produce high-quality embeddings independently, thus enhancing the model's overall performance. This approach contrasts sharply with the performance of larger models such as Llama 2 <ref type="bibr" target="#b19">[20]</ref>, which, despite having over 7 billion parameters, underperforms when a classifier head is added. This observation corroborates the notion that model size does not necessarily correlate with task-specific performance. Optimizing the model architecture to improve linguistic encoding, as demonstrated by mxbai-embed-large-v1, proves more beneficial than merely increasing the number of parameters.</p><p>Additionally, the performance of the Mamba 370m model, which matches mxbai-embed-large-v1 with an MCC of 0.793, highlights the potential of alternative architectures beyond Transformers. The Mamba model, with its state-space approach and selective propagation mechanism, presents a compelling case for further exploration of non-Transformer architectures in NLP tasks.</p><p>The horrible performance of the large language model Claude 2.0 Opus<ref type="foot" target="#foot_5">6</ref> in English, which is typically a benchmark model for complex reasoning tasks, warrants further investigation. Despite its state-of-the-art status in reasoning datasets, Claude 2.0 Opus showed a tendency to classify texts as conspiracy-related. This bias was evident even when the model provided reasoning for its classifications, suggesting a predisposition influenced by the sensitive nature of the subject matter. This finding highlights the need for ongoing research into model biases and their impact on classification tasks, particularly for topics with significant socio-cultural implications such as conspiracy theories.</p><p>In the Spanish dataset (see Table <ref type="table">3</ref>, the bert-base-spanish-wwm-uncased model achieved lower performance compared to its English counterparts, indicating potential limitations in the Spanish training data or model architecture. However, when using the multilingual model multilingual-e5-large, which was trained on Spanish data alone, still not surpassing the monolingual model. This suggests that in this context, monolingual models might be more effective than multilingual ones just using one language data for training. Interestingly, when the multilingual model was trained on both English and Spanish datasets, it achieved an MCC of 0.725, as shown in Table <ref type="table" target="#tab_2">4</ref>. This indicates that multilingual models can leverage larger and more diverse datasets to enhance their understanding of the task. The ability of multilingual models to generalize across languages is particularly evident when they are exposed to substantial amounts of well-represented data, demonstrating their potential to exploit linguistic diversity for improved performance.</p><p>Overall, we select the following models for the two trial of the competition:</p><p>• RUN 1 -Monolingual approach consists of Mamba 370m model for English and bert-base-spanishwwm-uncased model for Spanish. • RUN 2 -Multilingual approach consists of multilingual model multilingual-e5-large, trained in both languages together.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Parameters Tuning, Importance and Correlation</head><p>The results of our hyperparameter tuning highlight the significant influence of learning rate (lr) on model performance, with an importance score of 0.531 and a negative correlation of -0.535 with MCC. This suggests that optimizing the learning rate is crucial for achieving higher performance, as inappropriate values can lead to suboptimal results. Runtime also showed considerable importance (0.196) and a positive correlation (0.410), indicating that longer training times generally improve model performance.</p><p>Weight decay and sigmoid focal loss parameters, while less influential than the learning rate, still play vital roles. The weight decay parameter had an importance of 0.163 and a negative correlation of -0.323, suggesting that higher weight decay might adversely affect the model. Sigmoid focal loss <ref type="bibr" target="#b20">[21]</ref> parameters (alpha and gamma) demonstrated moderate importance, with alpha showing a positive correlation (0.156) and gamma a negative one (-0.325). The focal loss function, designed to address class imbalance, is given by:</p><formula xml:id="formula_0">FL(𝑝 𝑡 ) = −𝛼 𝑡 (1 − 𝑝 𝑡 ) 𝛾 log(𝑝 𝑡 )</formula><p>where 𝑝 𝑡 is the model's estimated probability for the true class label, 𝛼 𝑡 is a weighting factor for class imbalance, and 𝛾 is a focusing parameter that adjusts the rate at which easy examples are downweighted. This indicates a complex relationship where these parameters can be fine-tuned to balance the model's sensitivity to class imbalances effectively.</p><p>Other parameters, such as epochs, batch size, and accumulation steps, showed lower importance scores. Interestingly, batch size had a positive correlation with MCC (0.287), indicating that larger batch sizes might contribute to better performance. However, the relatively low importance scores for these parameters suggest that while they do influence performance, their impact is less critical compared to the learning rate and regularization parameters.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Official Competition Results and Conclusion</head><p>Table <ref type="table" target="#tab_3">5</ref> presents the test results, demonstrating that Run 2, which employs a single multilingual model, outperforms the monolingual models. Notably, in both languages, the baseline performance of BERT (MCC 0.7964) is exceeded. Out of 82 teams, only 17 surpass this threshold. Particularly striking is the performance in Spanish, where only 13 teams exceed the MCC threshold of 0.6681, placing us in the top three. These results highlight the potential advantages of multilingual models in achieving robust performance across languages. The improved performance in Run 2 suggests that training a multilingual model on data from both languages mitigates the so-called "curse of multilinguality", where multilingual models often struggle to distribute their knowledge equally across all languages. This phenomenon has been documented in the literature, where multilingual models tend to underperform on low-resource languages due to an imbalance in data distribution and representation <ref type="bibr" target="#b21">[22,</ref><ref type="bibr" target="#b22">23]</ref>. Our findings supports that providing a balanced dataset across languages can significantly enhance the fairness and effectiveness of multilingual models, as other works have applied these to counter content evasion on social media platforms <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b23">24]</ref>.</p><p>Furthermore, this study underscores the importance of addressing biases in large language models (LLMs). The bias observed in the Claude 2.0 Opus model, which showed a tendency to classify texts as conspiracy-related, raises critical questions about the ethical deployment of AI technologies. Such biases can have profound implications for freedom of expression and the equitable treatment of diverse perspectives. Future research should focus on developing techniques to identify and mitigate these biases, ensuring that LLMs operate fairly across different socio-cultural contexts.</p><p>In conclusion, our findings support the utility of multilingual models in handling diverse linguistic data, provided that training data is well-distributed across languages. The competition has allow us to conduct a research that demonstrate the potential of such models in achieving high performance and also emphasizes the necessity of continuous efforts to address and mitigate inherent biases in AI systems. Moving forward, it is essential to explore advanced methodologies for bias detection and mitigation, which will be crucial for the ethical and effective application of AI technologies in realworld scenarios.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Distribution of Text Length in Training Set by Category and Language</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Parameter Importance and Correlation with MCC performance</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Hyperparameters Explored in Bayesian Optimization</figDesc><table><row><cell>Parameter</cell><cell>Range/Values</cell><cell>Distribution/Method</cell></row><row><cell>Epochs</cell><cell>1 to 4</cell><cell>Uniform</cell></row><row><cell>Accumulation Steps</cell><cell>[1, 2, 3, 4]</cell><cell>Discrete</cell></row><row><cell>Learning Rate (lr)</cell><cell>1e-6 to 1e-4</cell><cell>Log-Uniform</cell></row><row><cell>Weight Decay</cell><cell>0.0 to 1.0</cell><cell>Uniform</cell></row><row><cell>Batch Size</cell><cell>[4]</cell><cell>Discrete</cell></row><row><cell>Loss Function</cell><cell>[BCEWithLogitsLoss, sigmoid_focal_loss]</cell><cell>Discrete</cell></row><row><cell>Sigmoid Focal Loss Alpha</cell><cell>0.25 to 1.0</cell><cell>Uniform</cell></row><row><cell>Sigmoid Focal Loss Gamma</cell><cell>1.0 to 4.0</cell><cell>Uniform</cell></row><row><cell>Scheduler Name</cell><cell>[PolynomialLR, CosineAnnealingWarmRestarts, LinearLR ConstantLR]</cell><cell>Discrete</cell></row><row><cell>Optimizer</cell><cell>[AdamW, Adam]</cell><cell>Discrete</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>Model performance metrics on validation set for English</figDesc><table><row><cell cols="2">Language Model</cell><cell cols="3">Accuracy F1-macro MCC</cell></row><row><cell></cell><cell>mamba-370m</cell><cell>0.908</cell><cell>0.861</cell><cell>0.793</cell></row><row><cell></cell><cell>mxbai-embed-large-v1</cell><cell>0.906</cell><cell>0.865</cell><cell>0.793</cell></row><row><cell></cell><cell>Baseline (BERT)</cell><cell>0.905</cell><cell>0.854</cell><cell>0.787</cell></row><row><cell></cell><cell>twitter-roberta-base-sentiment-latest</cell><cell>0.900</cell><cell>0.851</cell><cell>0.776</cell></row><row><cell>EN</cell><cell>DistilRoBERTa all-MiniLM-L6-v2_EN</cell><cell>0.896 0.896</cell><cell>0.856 0.851</cell><cell>0.776 0.771</cell></row><row><cell></cell><cell>DistilBERT</cell><cell>0.889</cell><cell>0.829</cell><cell>0.750</cell></row><row><cell></cell><cell>XLM-Roberta</cell><cell>0.869</cell><cell>0.826</cell><cell>0.729</cell></row><row><cell></cell><cell>nomic-embed-text-v1.5</cell><cell>0.870</cell><cell>0.782</cell><cell>0.710</cell></row><row><cell></cell><cell>LLAMA 2</cell><cell>0.803</cell><cell>0.730</cell><cell>0.578</cell></row><row><cell></cell><cell>Claude 2.0 Opus</cell><cell>0.457</cell><cell>0.366</cell><cell>0.236</cell></row><row><cell>Table 3</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="2">Model performance metrics on validation set for Spanish</cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="2">Language Model</cell><cell cols="3">Accuracy F1-macro MCC</cell></row><row><cell></cell><cell>bert-base-spanish-wwm-uncased</cell><cell>0.863</cell><cell>0.793</cell><cell>0.699</cell></row><row><cell></cell><cell>multilingual-e5-large</cell><cell>0.861</cell><cell>0.785</cell><cell>0.698</cell></row><row><cell>ES</cell><cell>mamba-370m nomic-embed-text-v1.5</cell><cell>0.843 0.838</cell><cell>0.770 0.763</cell><cell>0.654 0.643</cell></row><row><cell></cell><cell>XLM-Roberta</cell><cell>0.829</cell><cell>0.732</cell><cell>0.624</cell></row><row><cell></cell><cell>LLAMA 2</cell><cell>0.705</cell><cell>0.658</cell><cell>0.424</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 4</head><label>4</label><figDesc>Model performance metrics on validation set for English and Spanish</figDesc><table><row><cell cols="2">Language Model</cell><cell cols="6">EN Score Accuracy F1-macro MCC Accuracy F1-macro MCC ES Score</cell></row><row><cell>EN-ES</cell><cell>multilingual-e5-large multilingual-e5-base</cell><cell>0.896 0.895</cell><cell>0.845 0.850</cell><cell>0.768 0.769</cell><cell>0.874 0.868</cell><cell>0.821 0.819</cell><cell>0.725 0.714</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 5</head><label>5</label><figDesc>Comparison of Monolingual and Multilingual Approaches</figDesc><table><row><cell></cell><cell cols="4">Run 1 -Monolingual Approach Run 2 -Multilingual Approach</cell></row><row><cell></cell><cell>EN</cell><cell>ES</cell><cell>EN</cell><cell>ES</cell></row><row><cell>MCC</cell><cell>0.7894</cell><cell>0.6445</cell><cell>0.7965</cell><cell>0.7028</cell></row><row><cell>F1-macro</cell><cell>0.8947</cell><cell>0.8160</cell><cell>0.8977</cell><cell>0.8497</cell></row><row><cell cols="2">F1-conspiracy 0.8617</cell><cell>0.7523</cell><cell>0.8637</cell><cell>0.8035</cell></row><row><cell>F1-critical</cell><cell>0.9276</cell><cell>0.8796</cell><cell>0.9317</cell><cell>0.8960</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://huggingface.co/distilbert/distilbert-base-uncased</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">https://huggingface.co/spaces/mteb/leaderboard</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">https://huggingface.co/state-spaces/mamba-370m-hf</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4">https://wandb.ai/huertas_97/PAN_2024_Opposing/workspace</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_5">https://www.anthropic.com/news/claude-2</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">A Systematic Review and Meta-Analysis of Psychological Research on Conspiracy Beliefs: Field Characteristics, Measurement Instruments, and Associations With Personality Traits</title>
		<author>
			<persName><forename type="first">A</forename><surname>Goreis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Voracek</surname></persName>
		</author>
		<idno type="DOI">10.3389/fpsyg.2019.00205</idno>
		<ptr target="https://www.frontiersin.org/article/10.3389/fpsyg.2019.00205/full.doi:10.3389/fpsyg.2019.00205" />
	</analytic>
	<monogr>
		<title level="j">Frontiers in Psychology</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page">205</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Conspiratorial Beliefs and Cognitive Styles: An Integrated Look on Analytic Thinking, Critical Thinking, and Scientific Reasoning in Relation to (Dis)trust in Conspiracy Theories</title>
		<author>
			<persName><forename type="first">B</forename><surname>Gjoneska</surname></persName>
		</author>
		<idno type="DOI">10.3389/fpsyg.2021.736838</idno>
		<ptr target="https://www.frontiersin.org/articles/10.3389/fpsyg.2021.736838/full.doi:10.3389/fpsyg.2021.736838" />
	</analytic>
	<monogr>
		<title level="j">Frontiers in Psychology</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page">736838</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Factercheck: Semi-automated fact-checking through semantic similarity and natural language inference</title>
		<author>
			<persName><forename type="first">A</forename><surname>Martín</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Huertas-Tato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Álvaro</forename><surname>Huertas-García</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Villar-Rodríguez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Camacho</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.knosys.2022.109265</idno>
		<idno>doi:</idno>
		<ptr target="https://doi.org/10.1016/j.knosys.2022.109265" />
	</analytic>
	<monogr>
		<title level="j">Knowledge-Based Systems</title>
		<imprint>
			<biblScope unit="volume">251</biblScope>
			<biblScope unit="page">109265</biblScope>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Conspiracy or not? a deep learning approach to spot it on twitter</title>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">A</forename><surname>Galende</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Hernández-Peñaloza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Uribe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">A</forename><surname>García</surname></persName>
		</author>
		<idno type="DOI">10.1109/ACCESS.2022.3165226</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page" from="38370" to="38378" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Hamid</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shiekh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Said</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Ahmad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gul</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Hassan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Al-Fuqaha</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2012.07517</idno>
		<title level="m">Fake news detection in social media using graph neural networks and nlp techniques: A covid-19 use-case</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Countering malicious content moderation evasion in online social networks: Simulation and detection of word camouflage</title>
		<author>
			<persName><forename type="first">Á</forename><surname>Huertas-García</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Martín</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Huertas-Tato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Camacho</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.asoc.2023.110552</idno>
		<idno>.110552</idno>
		<ptr target="https://doi.org/10.1016/j.asoc.2023" />
	</analytic>
	<monogr>
		<title level="j">Applied Soft Computing</title>
		<imprint>
			<biblScope unit="volume">145</biblScope>
			<biblScope unit="page">110552</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Nourishing critical thinking skills using neuro-linguistic programming: farah hashmi</title>
		<author>
			<persName><forename type="first">Editor</forename><surname>Pje</surname></persName>
		</author>
		<idno type="DOI">10.30971/pje.v39i1.865</idno>
		<ptr target="https://ojs.aiou.edu.pk/index.php/pje/article/view/865.doi:10.30971/pje.v39i1.865" />
	</analytic>
	<monogr>
		<title level="j">PJE</title>
		<imprint>
			<biblScope unit="volume">39</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<author>
			<persName><forename type="first">H</forename><surname>Devinney</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Björklund</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Björklund</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2205.02526</idno>
		<title level="m">Theories of &quot;gender&quot; in nlp bias research</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Overview of PAN 2024: Multi-Author Writing Style Analysis, Multilingual Text Detoxification, Oppositional Thinking Analysis, and Generative AI Authorship Verification</title>
		<author>
			<persName><forename type="first">J</forename><surname>Bevendorff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><forename type="middle">B</forename><surname>Casals</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chulvi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dementieva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Elnagar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Freitag</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Korenčić</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mayerl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mukherjee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Rangel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Smirnova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Stamatatos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Taulé</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Ustalov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wiegmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Zangerle</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fourteenth International Conference of the CLEF Association (CLEF 2024)</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<meeting><address><addrLine>Berlin Heidelberg New York</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Overview of pan 2024: multi-author writing style analysis, multilingual text detoxification, oppositional thinking analysis, and generative ai authorship verification</title>
		<author>
			<persName><forename type="first">J</forename><surname>Bevendorff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><forename type="middle">B</forename><surname>Casals</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chulvi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dementieva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Elnagar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Freitag</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Korenčić</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mayerl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mukherjee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">European Conference on Information Retrieval</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2024">2024</date>
			<biblScope unit="page" from="3" to="10" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Attention is all you need</title>
		<author>
			<persName><forename type="first">A</forename><surname>Vaswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shazeer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Parmar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Uszkoreit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">N</forename><surname>Gomez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ł</forename><surname>Kaiser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Polosukhin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in neural information processing systems</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Mamba: Linear-time sequence modeling with selective state spaces</title>
		<author>
			<persName><forename type="first">A</forename><surname>Gu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Dao</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2312.00752</idno>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">BERT: pre-training of deep bidirectional transformers for language understanding</title>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
		<idno>CoRR abs/1810.04805</idno>
		<ptr target="http://arxiv.org/abs/1810.04805.arXiv:1810.04805" />
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<author>
			<persName><forename type="first">Z</forename><surname>Nussbaum</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">X</forename><surname>Morris</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Duderstadt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mulyar</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2402.01613</idno>
		<title level="m">Nomic embed: Training a reproducible long context text embedder</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<title level="m" type="main">Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter</title>
		<author>
			<persName><forename type="first">V</forename><surname>Sanh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Debut</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chaumond</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Wolf</surname></persName>
		</author>
		<idno>ArXiv abs/1910.01108</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">TweetNLP: Cutting-edge natural language processing for social media</title>
		<author>
			<persName><forename type="first">J</forename><surname>Camacho-Collados</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Rezaee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Riahi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ushio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Loureiro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Antypas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Boisson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Espinosa Anke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">Martínez</forename><surname>Cámara</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2022.emnlp-demos.5</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing: System Demonstrations</title>
				<editor>
			<persName><forename type="first">W</forename><surname>Che</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><surname>Shutova</surname></persName>
		</editor>
		<meeting>the 2022 Conference on Empirical Methods in Natural Language Processing: System Demonstrations<address><addrLine>Abu Dhabi, UAE</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="38" to="49" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Shakir</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Koenig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lipp</surname></persName>
		</author>
		<ptr target="https://www.mixedbread.ai/blog/mxbai-embed-large-v1" />
		<title level="m">Open Source Strikes Bread -New Fluffy Embedding Model</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Kusupati</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Bhatt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rege</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wallingford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Sinha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Ramanujan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Howard-Snyder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kakade</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Jain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Farhadi</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2205.13147</idno>
		<title level="m">Matryoshka representation learning</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Spanish pre-trained bert model and evaluation data</title>
		<author>
			<persName><forename type="first">J</forename><surname>Cañete</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Chaperon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Fuentes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J.-H</forename><surname>Ho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Kang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Pérez</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">PML4DC at ICLR</title>
				<imprint>
			<date type="published" when="2020">2020. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<author>
			<persName><forename type="first">H</forename><surname>Touvron</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Martin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Stone</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Albert</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Almahairi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Babaei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Bashlykov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Batra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Bhargava</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Bhosale</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Bikel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Blecher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">C</forename><surname>Ferrer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Cucurull</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Esiobu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Fernandes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Fu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Fu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Fuller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Goswami</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hartshorn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Hosseini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Hou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Inan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kardas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Kerkez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Khabsa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Kloumann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Korenev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">S</forename><surname>Koura</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-A</forename><surname>Lachaux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lavril</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Liskovich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Lu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Mao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Martinet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Mihaylov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Mishra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Molybog</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Nie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Poulton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Reizenstein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Rungta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Saladi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Schelten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Silva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">M</forename><surname>Smith</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Subramanian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><forename type="middle">E</forename><surname>Tan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Taylor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Williams</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">X</forename><surname>Kuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Yan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Zarov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Fan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kambadur</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Narang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rodriguez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Stojnic</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Edunov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Scialom</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2307.09288</idno>
		<title level="m">Llama 2: Open foundation and fine-tuned chat models</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<title level="m" type="main">Focal loss for dense object detection</title>
		<author>
			<persName><forename type="first">T.-Y</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Girshick</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Dollár</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1708.02002</idno>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Conneau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Khandelwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Chaudhary</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Wenzek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Guzmán</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Grave</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ott</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Stoyanov</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1911.02116</idno>
		<title level="m">Unsupervised cross-lingual representation learning at scale</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<title level="m" type="main">Xtreme: A massively multilingual multi-task benchmark for evaluating cross-lingual generalization</title>
		<author>
			<persName><forename type="first">J</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ruder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Siddhant</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Neubig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Firat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Johnson</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2003.11080</idno>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Countering Misinformation Through Semantic-Aware Multilingual Models</title>
		<author>
			<persName><forename type="first">Á</forename><surname>Huertas-García</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Martín</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Huertas-Tato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Camacho</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-91608-4_31</idno>
	</analytic>
	<monogr>
		<title level="m">Intelligent Data Engineering and Automated Learning -IDEAL 2021</title>
				<editor>
			<persName><forename type="first">H</forename><surname>Yin</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Camacho</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Tino</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Allmendinger</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">J</forename><surname>Tallón-Ballesteros</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">K</forename><surname>Tang</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S.-B</forename><surname>Cho</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Novais</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Nascimento</surname></persName>
		</editor>
		<meeting><address><addrLine>Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer International Publishing</publisher>
			<date type="published" when="2021">2021</date>
			<biblScope unit="volume">13113</biblScope>
			<biblScope unit="page" from="312" to="323" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
