<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Hierocles of Alexandria at Touché: Multi-task &amp; Multi-head Custom Architecture with Transformer-based Models for Human Value Detection Notebook for the Touché Lab at CLEF 2024</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Sotirios</forename><surname>Legkas</surname></persName>
							<email>sotirislegkas@iit.demokritos.gr</email>
							<affiliation key="aff0">
								<orgName type="department">Institute of Informatics &amp; Telecommunications</orgName>
								<orgName type="institution">National Centre for Scientific Research (N.C.S.R.) &apos;Demokritos&apos;</orgName>
								<address>
									<addrLine>Aghia Paraskevi</addrLine>
									<settlement>Attica</settlement>
									<country key="GR">Greece</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Christina</forename><surname>Christodoulou</surname></persName>
							<email>ch.christodoulou@iit.demokritos.gr</email>
							<affiliation key="aff0">
								<orgName type="department">Institute of Informatics &amp; Telecommunications</orgName>
								<orgName type="institution">National Centre for Scientific Research (N.C.S.R.) &apos;Demokritos&apos;</orgName>
								<address>
									<addrLine>Aghia Paraskevi</addrLine>
									<settlement>Attica</settlement>
									<country key="GR">Greece</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Matthaios</forename><surname>Zidianakis</surname></persName>
							<email>mzidianakis@iit.demokritos.gr</email>
							<affiliation key="aff0">
								<orgName type="department">Institute of Informatics &amp; Telecommunications</orgName>
								<orgName type="institution">National Centre for Scientific Research (N.C.S.R.) &apos;Demokritos&apos;</orgName>
								<address>
									<addrLine>Aghia Paraskevi</addrLine>
									<settlement>Attica</settlement>
									<country key="GR">Greece</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Dimitrios</forename><surname>Koutrintzes</surname></persName>
							<email>dkoutrintzes@iit.demokritos.gr</email>
							<affiliation key="aff0">
								<orgName type="department">Institute of Informatics &amp; Telecommunications</orgName>
								<orgName type="institution">National Centre for Scientific Research (N.C.S.R.) &apos;Demokritos&apos;</orgName>
								<address>
									<addrLine>Aghia Paraskevi</addrLine>
									<settlement>Attica</settlement>
									<country key="GR">Greece</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Maria</forename><surname>Dagioglou</surname></persName>
							<email>mdagiogl@iit.demokritos.gr</email>
							<affiliation key="aff0">
								<orgName type="department">Institute of Informatics &amp; Telecommunications</orgName>
								<orgName type="institution">National Centre for Scientific Research (N.C.S.R.) &apos;Demokritos&apos;</orgName>
								<address>
									<addrLine>Aghia Paraskevi</addrLine>
									<settlement>Attica</settlement>
									<country key="GR">Greece</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Georgios</forename><surname>Petasis</surname></persName>
							<email>petasis@iit.demokritos.gr</email>
							<affiliation key="aff0">
								<orgName type="department">Institute of Informatics &amp; Telecommunications</orgName>
								<orgName type="institution">National Centre for Scientific Research (N.C.S.R.) &apos;Demokritos&apos;</orgName>
								<address>
									<addrLine>Aghia Paraskevi</addrLine>
									<settlement>Attica</settlement>
									<country key="GR">Greece</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Hierocles of Alexandria at Touché: Multi-task &amp; Multi-head Custom Architecture with Transformer-based Models for Human Value Detection Notebook for the Touché Lab at CLEF 2024</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">E129B357A8D7704BF4071C058C630D71</idno>
					<idno type="arXiv">arXiv:2109.04712.</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:02+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>human values, multi-label text classification, custom multi-head architecture, multi-lingual, transformers (G. Petasis) 0009-0000-9468-5650 (S. Legkas)</term>
					<term>0009-0009-5616-3937 (C. Christodoulou)</term>
					<term>0009-0009-5102-8103 (M. Zidianakis)</term>
					<term>0009-0003-7401-6347 (D. Koutrintzes)</term>
					<term>0000-0002-3357-2844 (M. Dagioglou)</term>
					<term>0000-0003-3157-1597 (G. Petasis)</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The paper presents our participation as Hierocles of Alexandria in Touché at CLEF 2024, which addressed the Human Value Detection shared task. The objectives of the task was to detect one or more human values (sub-task 1) and their attainment (sub-task 2) in lengthy texts across nine languages, including the automatic translation of these texts into English. Our methodology involved the fine-tuning of four Transformer language models within a customized multi-head model architecture for multi-label text classification. The experimental approach comprised comprehensive data analysis, the utilization of various loss functions, and class positive weights to handle class imbalance. Additionally, we incorporated previous sentences as context and represented human values as special tokens in the texts to enhance classification performance. Notably, all our submissions for the multi-lingual data surpassed the baseline submissions in both sub-tasks 1 and 2. Our top-performing submission secured the 1 𝑠𝑡 position among all the participating teams in sub-task 1 in both the multi-lingual and English-translated data.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Values motivate our actions <ref type="bibr" target="#b0">[1]</ref> and impact all processes of our (moral) behaviour from perception and judgment to focus and action <ref type="bibr" target="#b1">[2]</ref>. Being essentially the driving forces of individuals and societies, intelligibly identifying them empowers us to understand more profoundly, among others, our cultural heritage <ref type="bibr" target="#b2">[3]</ref>, citizen's political behaviour <ref type="bibr" target="#b3">[4]</ref> and human interaction with artificial agents <ref type="bibr" target="#b4">[5]</ref>. This knowledge can be fed back to people through the delivery of sustainable and responsible solutions from the related duty holders. Naturally, narratives are vessels of values. Historical texts, social media content, news items, and ChatGPT products are all resources to extract values and inform research, resolve sociopolitical tensions, deliver responsible AI. In particular, Touché <ref type="bibr" target="#b5">[6]</ref> aims to advance our understanding of decision-making and opinion-forming processes by supporting the development of related methods and tools based on human values detection.</p><p>Human values detection in natural language is a complex task due to diverse perceptions, multilingualism, terminology interpretation, values attainment and actor attribution, among others. These are challenges that we have encountered through our research <ref type="bibr" target="#b6">[7,</ref><ref type="bibr" target="#b7">8]</ref> and our participation in relevant projects 1,2 , and are also reflected in the performance of the models developed as part of SemEval-2023 Task 4: ValueEval <ref type="bibr" target="#b8">[9]</ref>. Touché <ref type="bibr" target="#b5">[6]</ref> at CLEF 2024 provides the opportunity to examine many of the aforementioned challenges. The provided dataset is a collection of 3000 human-annotated texts, including news articles and political texts, chosen to reflect diverse views. Over 70 people, from 9 language teams, annotated texts (in their mother tongue) for their value content and attainment. Schwartz's values <ref type="bibr" target="#b0">[1]</ref> were adopted for the annotation vocabulary. In addition to the original files provided in nine languages, a machine-translated version of them in English was provided.</p><p>The current state of the art in Natural Language Processing (NLP) and ML/AI has enabled the development of methods that identify human values in natural language artifacts <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b10">11,</ref><ref type="bibr" target="#b11">12,</ref><ref type="bibr" target="#b12">13]</ref>. Advanced techniques for text classification, particularly for shorter text sequences, rely on fine-tuning Transformer-based models <ref type="bibr" target="#b6">[7]</ref>, Large Language Models (LLMs) <ref type="bibr" target="#b13">[14]</ref>, ensembles of Transformer-based models <ref type="bibr" target="#b14">[15,</ref><ref type="bibr" target="#b15">16]</ref>, and custom model architectures involving multiple heads with attention mechanisms <ref type="bibr" target="#b16">[17,</ref><ref type="bibr" target="#b17">18]</ref>.</p><p>This paper presents the methodology and results of Hierocles of Alexandria team in the Human Value Detection shared task of the second edition of Touché <ref type="bibr" target="#b5">[6]</ref> at CLEF. Motivated by the availability of longer texts in this Touché edition and the availability of multilingual data, the innovative aspects of our approach include the modeling of contextual information, and the application of multi-task learning. Assuming that the value classification of a sentence may depend on earlier sentences and their classifications, previous sentences and their labels (either from annotated data during training or from classification of previous sentences during evaluation) are provided as input along with the sentence under classification. Multi-task learning, in the form of language-specific classification tasks, has been employed in order to capture potential different value instantiations in different languages.</p><p>Our approach leveraged fine-tuning four Transformer-based language models within a custom multitask with multiple heads model architecture, specifically tailored for multi-lingual and multi-label text classification in order to capture the linguistic nuances. Our experimental strategy comprised a comprehensive data analysis, the application of various loss functions, and the utilization of class positive weights to mitigate the challenge of class imbalance. Our approach achieved the highest score, securing the 1 𝑠𝑡 place for both multilingual and English submissions in sub-task 1, surpassing all other participating teams and baselines. The code for our approach is available on the provided GitHub link. <ref type="foot" target="#foot_0">3</ref>In the context of Touché, the results from all approaches were submitted through the TIRA platform, which ensured the reproducibility and reliability of the software employed by participants, thereby facilitating the comparison of information retrieval experiments <ref type="bibr" target="#b18">[19]</ref>.</p><p>The structure of this paper is as follows: Section 2 explores the background. Several aspects of the data, including data analysis, pre-processing and an exploratory phase, are presented in Section 3. Section 4 introduces an overview of the developed system and the experiments. Section 5 presents the results. Finally, in Section 6 the conclusions are discussed, including limitations and future work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Background</head><p>The exploration of human values within Natural Language Processing (NLP) encompasses various theoretical and empirical endeavors. Central to this exploration is Shalom H. Schwartz's theory of basic human values, which identifies nineteen universal values inherent to human behavior and cultural expression. These values, driven by distinct motivational goals, form a circular structure that illustrates their dynamic interplay, where pursuing one value may align with or conflict with another <ref type="bibr" target="#b0">[1]</ref>. Schwartz's framework provides valuable insights into the motivational goals driving human actions and the complex interrelations among different values. In NLP, this framework offers a robust foundation for identifying and interpreting human values embedded within language.</p><p>In a significant research endeavor focused on identifying human values in NLP, a comprehensive taxonomy comprising 54 human values was crafted, aligning closely with psychological research. The researchers also introduced the initial annotated dataset for studying human values behind arguments <ref type="bibr" target="#b19">[20]</ref>. This dataset encompassed 5.270 arguments from four distinct cultures: Africa, China, India, and the USA. Each argument in the dataset consisted of a premise, a conclusion, and a stance attribute indicating whether the premise supported or opposed the conclusion. The researchers manually annotated these arguments for human values. Their methodology has paved the way for automating the classification of human values, yielding promising results, with F1-scores reaching up to 0,81 and averaging 0,25, establishing a benchmark for future research in this domain.</p><p>To further advance the field of human values detection in argumentative texts, the authors of the aforementioned research organized the ValueEval: Identification of Human Values Behind Arguments shared task 4 in SemEval-2023 <ref type="bibr" target="#b8">[9]</ref> by mapping the 54 human values from their previous research to a set of 20 value categories for multi-label classification. The task showcased both the potential and challenges associated with identifying human values in argumentative texts. A total of 39 teams contributed their methodologies, utilizing the Touché23-ValueEval Dataset comprising 9.324 arguments sourced from 6 diverse outlets, including religious texts, political forums, free-text arguments, newspaper editorials, and online democracy platforms in English <ref type="bibr" target="#b20">[21]</ref>. Each argument included a premise, a conclusion, and a stance attribute signifying whether the premise was in favor of or against the conclusion. The teams' approaches were primarily evaluated on the Macro-F1 score.</p><p>The task's winner, the Adam-Smith team, achieved an F1 score of 0,56 by calculating a global decision threshold during training that optimizes the F1 score. They mainly employed twelve individual Transformer-based models that are ensembled in order to perform multi-label classification. <ref type="bibr" target="#b15">[16]</ref>. The second-place John-Arthur team found that it is beneficial to encode the input data by adding tokenizer's special token separators, corresponding to low-cardinality values of Stance (in favour of vs against). Also, they fine-tuned larger Language Models, which performed better. Lastly, they adopted a threshold of 0,2 at the output of the sigmoid function to get the binary predictions for each human value, achieving an F1 score of 0,55 <ref type="bibr" target="#b13">[14]</ref>. Addressing implicit value discrimination and data imbalance, the PAI team employed a multi-label classification model with a class-balanced loss function, securing multiple top positions across task categories with an overall average score of 0,54, placing them third <ref type="bibr" target="#b14">[15]</ref>. The Mao-Zedong team's introduction of a multi-head attention mechanism and a contrastive learningenhanced K-nearest neighbor mechanism resulted in an F1 score of 0,53, placing them fourth <ref type="bibr" target="#b16">[17]</ref>. Finally, certain members of the Hierocles of Alexandria team, who participated in that year's task as part of the Andronicus of Rhodes team, leveraged a Transformer model with four classification heads and applied two classification strategies with different activation and loss functions. In addition, they used two different data partitioning methods to handle class imbalance. Their system, employing majority voting, achieved an F1 score of 0,48, placing them in the upper half of the competition <ref type="bibr" target="#b6">[7]</ref>.</p><p>Inspired by the best methodologies employed in ValueEval, our approach aimed to tackle class imbalance and improve the classification performance, through the use of sigmoid threshold, larger language models, and tokenizer's special tokens in the encoded input. Nevertheless, Touché's Human Value Detection task at CLEF 2024 <ref type="bibr" target="#b5">[6]</ref> has extended human value detection by integrating multiple languages, besides English. The introduction of new languages introduces more challenges, such as possible differences in annotation styles among languages, which adds complexity to the problem. To this end, our proposed approach was customised to the dataset features and the task by incorporating techniques that address the issue of multi-linguality and capture the linguistic nuances. The introduction of multiple languages and the need to address language-specific phenomena, was the main motivation behind our proposed approach in this paper, which includes a model architecture with multiple heads that are specifically tailored to each language, aiming to model more accurately multi-label text classification across multiple languages.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Data</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Data Analysis</head><p>The dataset comprises 2.648 complete texts in nine languages: English, Greek, German, French, Bulgarian, Hebrew, Italian, Dutch, and Turkish. The dataset is split by the shared task organisers into training, validation and test sets: Of these texts, 1.603 are used for training, 523 for validation, and 522 for testing.</p><p>The number of annotated texts per language varies, as illustrated in Table <ref type="table" target="#tab_0">1</ref>. English has the highest number of texts (408), while French and Hebrew have the fewest (219 and 250, respectively). Each text is segmented into sentences, resulting in 74.231 sentences: 44.758 for training, 14.904 for validation, and 14.569 for testing.</p><p>The number of labels varies for each language, as shown in Table <ref type="table" target="#tab_0">1</ref>. There is no correlation between the number of texts and the number of labels. For instance, even though Hebrew texts were among the fewest, they had the highest number of labels (4.992).</p><p>For sub-task 1, the texts are annotated with Schwartz's 19 personal values <ref type="bibr" target="#b0">[1]</ref>. Nearly half of the sentences (30.662 out of 59.662) are labelled with one or more values. Labels for the test set are not provided to evaluate participating systems. For sub-task 2, each classified value includes an annotation indicating whether the value is attained or constrained by the sentence, resulting in a final dataset with 38 classes.</p><p>The frequency of human value labels varies quite a bit, as depicted in Figure <ref type="figure">1</ref>. Security: societal is the most frequently used value, with over 5.000 labels. Achievement and Conformity: rules are also quite popular, with over 3.000 labels each. On the other hand, Self-direction: thought, Universalism: tolerance, and Humility are less commonly used, with fewer than 1.000 labels. In fact, Humility is the least represented value, with only 151 labels.</p><p>All data was provided in the original language and translated using the DeepL API, except for Hebrew, which was translated using the Google Translate API. Figure <ref type="figure">1</ref>: Labels' frequency across all texts. Some labels like "Security societal", "Conformity: Interpersonal" and "Achievement" are more frequent within the dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Data Pre-processing</head><p>The dataset provides unique identifiers for each sentence, indicating the source text and its position within that text. Despite individual sentences being labeled independently, annotators were presented with the complete text for annotation, potentially influencing their assessments based on the overarching context.</p><p>To address this consideration, our pre-processing approach focuses on adding contextual information by integrating the previous part of the text and its annotated values. This was achieved through two main strategies: incorporating previous sentences and adding special tokens.</p><p>1. Incorporating Previous Sentences. For each sentence, we appended the two preceding sentences to each target sentence, thereby providing context from the specific text. If the total number of tokens exceeded the maximum allowed by our base model (maximum: 512), tokens were removed starting from the most distant sentence. If the sentence was the first sentence of a text, no preceding sentences were added. The "&lt;/s&gt;" separator token linked the preceding and target sentences. 2. Adding Special Tokens. We implemented special tokens to represent each class, such as "&lt;Security:</p><p>societal&gt;" representing the value "Security: societal". We used the annotated labels from the previous two sentences for each sentence and appended them to the end. No special token was added if there was no annotated label for the previous sentences. This enables the classifier to interpret the annotator's perspective for better contextual understanding. These tokens were added as special tokens in the model's tokenizer and the token embedding matrix of the model was resized. They were assigned to attributes in the tokenizer for easy access and to make sure they were not split during tokenization. For predictions on the validation and test sets, the predicted classes were used as special tokens to enhance the model's contextual understanding.</p><p>The following is an example of pre-processed English text for the model input:</p><p>[CLS] Having spoken to many different left-leaning Hispanics, Avila said, "they are really beginning to feel like the Democratic party has become too extreme to the point where it's starting to scare some of them." &lt;Security: societal&gt; &lt;/s&gt; Many are beginning to turn away from the Democratic party because "they're getting vibes of a communist Cuba and socialist Venezuela here in America. " As a result, Avila said Hispanics are going to be "extremely instrumental" in the upcoming midterm elections. &lt;Self-direction: action&gt; &lt;/s&gt; "They are starting to come to the realization that their conservative values are in opposition to what the media has been trying to feed them in favor of Biden and the Democrats. " [SEP]</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Exploratory Phase</head><p>We carried out several experiments to explore the behaviour of the pre-trained language models in order to form a baseline for our development process. This phase primarily exploits the language models to assess how well the collective human values of the mentioned dataset are captured by pretrained Transformer models <ref type="bibr" target="#b17">[18]</ref>, enhanced with classification heads so as to perform multi-label text classification, given the respective textual inputs. This process ensures that the models are fine-tuned to adequately fit the dataset with respect to all the language and sentence constraints.</p><p>The baseline experiments involved both multi-lingual (all languages together) and mono-lingual (each language separately) tests in order to record the effect of the special traits of each language on the human values. In general, we observed that the multi-lingual performance of the baseline models on the human value classification is higher than the performance on the individual languages in the mono-lingual experiments, as shown in Table <ref type="table">4</ref> of Appendix A. This outcome could be explained by the close relation of the several inherent features (e.g context, vocabulary) of each language to the human value perception.</p><p>To further inspect this correlation and to note the bias of each language, we developed a more specific architecture that is built upon the baseline models, in order to improve the performance.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">System Overview</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Model Architecture</head><p>Based on the findings in Section 3.1, Figure <ref type="figure">1</ref> illustrates the imbalance in label distribution across different languages. This imbalance is partly due to the varying annotation styles among languages. For instance, some languages, such as English, have a large number of examples but fewer annotations, while others, like Hebrew, have fewer examples but many annotations. Consequently, similar sentences in different languages may receive different annotations. Experimental results indicate that a single out-of-the-box pre-trained Transformer model fails to effectively capture the unique linguistic features of each language, given the different annotation styles per language. In contrast, models that are fine-tuned for a specific language outperform those that are fine-tuned across all languages.</p><p>To address the multi-lingual nature of the problem and the differences in annotations between each language, a custom ensemble model was constructed. The architecture, as seen in Figure <ref type="figure">2</ref>, leverages a pre-trained Transformer language model as its foundation. On top of this, nine custom Transformer heads were added, each tailored to a specific language: English, Greek, Dutch, Turkish, French, Bulgarian, Hebrew, Italian, and German.</p><formula xml:id="formula_0">Pretrained Base Model [CLS] 𝑇 −2 1 ... 𝑇 −2 𝑘 [𝑆𝑇 −2 ] 𝑇 −1 1 ... 𝑇 −1 𝑙 [𝑆𝑇 −1 ] 𝑇 1 𝑇 2 ... 𝑇 𝑚 [SEP]</formula><p>Language Splitter 1. The input batch is fed into the pre-trained base model (Transformer).</p><note type="other">Language</note><p>2. The output of the pre-trained model is passed through the language splitter which splits it according to the language identifiers within the batch. Each split tensor is directed to the corresponding custom Transformer head based on its language for further processing. 3. The logits produced by each custom Transformer head are concatenated into a single batch through the language combiner. 4. The concatenated logits batch is passed through the loss function to compute the training loss.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Model performs backpropagation.</head><p>This approach allows the model to handle multiple languages effectively by utilizing specialized components tailored to the linguistic features and annotation styles of each language.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Experimental Setup</head><p>The decision to utilize open-source Transformer-based multi-lingual models, specifically obtained from the Hugging Face platform, in our research was motivated by the multi-lingual composition of the data, which encompassed nine distinct languages. These models have been pre-trained across various languages, rendering them an optimal choice for providing a robust and comprehensive framework for analyzing and interpreting multi-lingual data. Consequently, we employed such models to ensure the effective capture of language-specific nuances and contexts, leading to more accurate and reliable results. For the data that underwent automatic translation into English, we employed open-source Transformer-based models that were exclusively pre-trained in English. This approach ensured an optimal understanding and interpretation of the nuances of the English language, thereby bolstering the accuracy of our analysis.</p><p>We utilized the multi-lingual base version of the RoBERTa Transformer-based language model <ref type="bibr" target="#b21">[22]</ref>, XLM-RoBERTa-base<ref type="foot" target="#foot_1">4</ref>  <ref type="bibr" target="#b22">[23]</ref>, which underwent pre-training on 100 languages with 768 layers. This model was employed to conduct preliminary experiments for multi-label text classification using AutoModelForSequenceClassification. Subsequently, baseline scores were obtained during the exploratory phase (See section 3.3). After analyzing the baseline results for individual languages and all languages collectively (see Table <ref type="table">4</ref> in Appendix A), we leveraged the larger 1024-layer version, XLM-RoBERTalarge<ref type="foot" target="#foot_2">5</ref>  <ref type="bibr" target="#b22">[23]</ref>, to conduct further experiments involving loss functions, class weights, and different class thresholds (see sections 4.3 and 4.4). The purpose was to address the challenges of class imbalance and language disparities. These experiments were primarily facilitated using the Transformers and Hugging Face libraries, in conjunction with 2 NVIDIA TITAN RTX GPU cards, with 24GB VRAM each.</p><p>For both sub-tasks, we fine-tuned two Transformer-based language models for the multi-lingual data using the custom model architecture with multiple heads presented in section 4.1, with each head focusing on a specific language. The employed models were the XLM-RoBERTa-large <ref type="bibr" target="#b22">[23]</ref> and the XLM-RoBERTa-xl<ref type="foot" target="#foot_3">6</ref>  <ref type="bibr" target="#b23">[24]</ref>, with 1024 and 2560 layers, respectively. In the case of the English-translated data, we integrated the RoBERTa-large<ref type="foot" target="#foot_4">7</ref>  <ref type="bibr" target="#b21">[22]</ref> and the DeBERTa-v2-xxl<ref type="foot" target="#foot_5">8</ref>  <ref type="bibr" target="#b24">[25]</ref> models, consisting of 1024 and 1536 layers, respectively, into the custom multi-head architecture, focusing solely on English for both sub-tasks.</p><p>We initially fine-tuned our models using the provided training data and fine-tuned them using the validation set. During the fine-tuning process, we established the hyperparameters, finalized the loss function, and determined the best thresholds for our submitted results. Then, we combined the training and validation data to use as the training set for fine-tuning, without having a separate validation set, using the previously defined hyperparameters. An overview of the hyperparameters used for our experiments and submissions is provided in Table <ref type="table" target="#tab_5">5</ref> of the Appendix A.</p><p>As for the custom model architecture, the custom head for RoBERTa-large and XLM-RoBERTa-large included three Transformer layers, while the custom head for XLM-RoBERTa-xl employed only one. Due to GPU VRAM memory and time limitations, the DeBERTa-v2-xxl did not incorporate any Transformer layers in its custom head. The experiments with the custom model architecture, which form the final submissions, were conducted using 2 NVIDIA H100 PCIe GPU cards, with 80GB VRAM each.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Loss Functions &amp; Class Weights</head><p>Various loss functions, including Binary Cross-Entropy Loss with Logits<ref type="foot" target="#foot_6">9</ref> , Focal Loss <ref type="bibr" target="#b25">[26]</ref>, Classbalanced Loss <ref type="bibr" target="#b26">[27]</ref>, Distribution-Balanced Loss <ref type="bibr" target="#b27">[28]</ref>, and Class-balanced Negative Tolerant Regularization Loss <ref type="bibr" target="#b28">[29]</ref>, were tested by modifying the Trainer class from Hugging Face. These loss functions were employed as they were originally developed for handling data imbalance issues. They have previously been employed for the detection of human values by the PAI team in SemEval-2023 <ref type="bibr" target="#b14">[15]</ref>. Positive weights were also calculated for each class to give more importance to the under-represented classes during model training, thereby improving the model's performance in these classes. The experiments using the XLM-RoBERTa-large with the standard classification head (AutoModelForSequenceClassification) showed that the Binary Cross-Entropy Loss with Logits achieved the best results. Therefore, this loss function was used for all the submitted runs with and without class positive weights.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4.">Thresholds</head><p>Initial experiments were conducted with various thresholds ranging from 0,1 to 0,95. The Macro-F1 score for all classes and each class separately was calculated during fine-tuning and evaluating with the provided validation set. After applying the sigmoid function to the validation and test set predictions, the predictions were converted into 1 if they were equal to or higher than the threshold and 0 if they were lower than the threshold. Consequently, 3 separate prediction files were created based on the 0,5 default threshold, the best general threshold for all classes, and the best threshold for each class. Based on the results from the validation set, the prediction file utilizing the optimal threshold for each class demonstrated the highest scores. Therefore, all predictions submitted for the test set were generated by determining the optimal threshold for each class individually.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Results</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Sub-task 1</head><p>The data presented in Table <ref type="table" target="#tab_1">2</ref> illustrates that our test set submissions demonstrated significant improvement over the baseline scores in both the multi-lingual and translated English datasets. Among the multi-lingual multi-label multi-head models, the XLM-RoBERTa-xl model achieved the highest Macro-F1 score (39%) across all 38 classes by utilizing context and special tokens without class positive weights and being fine-tuned on the combined training and validation data as the training set. Conversely, the XLM-RoBERTa-large model, employing context, special tokens, class positive weights, and fine-tuned on 19 classes using only the training data, achieved the lowest score (34%).</p><p>In the context of the translated English data, the XLM-RoBERTa-large model, utilizing context and special tokens without class positive weights and having been fine-tuned on the combined training and validation data as the training set, produced the lowest Macro-F1 score across all classes (35%). At the same time, the remaining submissions yielded identical scores (37%).</p><p>Upon examining the F1 scores for each class individually, it becomes apparent that the Universalism: nature class achieved the highest F1 score at 63%, signifying successful detection by the models, as the remaining scores do not fall below 59%. Conversely, the classes with lower frequency in the texts were less accurately detected by the models. For instance, values such as Humility received a 0% F1 score in most submissions, with the highest score reaching only 11%. Furthermore, the models struggled to accurately classify the Self-direction: thought value, as their scores remained below 20%. Despite being one of the minority classes, the models correctly detected at least 27% of the annotated labels in the Universalism: tolerance class. The different model performance in classes is also evident in Figure <ref type="figure" target="#fig_1">3</ref> of the Appendix A, which illustrates the radar plot of the 19 values through the performance of our top-performing XLM-RoBERTa-xl model compared to the baseline models. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">Sub-task 2</head><p>The data presented in Table <ref type="table" target="#tab_3">3</ref> illustrates that our submission for the multi-lingual test dataset outperformed the baseline score. Utilizing the XLM-RoBERTa-xl for the multi-lingual dataset and the </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion &amp; Future Work</head><p>Our system, developed for Touché at CLEF 2024, addressed the "Human Value Detection" shared task by participating in both sub-tasks. This involved the fine-tuning of four Transformer language models within a custom multi-head model architecture for multi-label text classification. Our experimental approach encompassed the utilization of loss functions and class positive weights, as well as the incorporation of previous sentences as context and the representation of human values as special tokens. These measures were implemented to mitigate class imbalance and enhance the models' capacity to comprehend and classify texts more effectively. Our submissions demonstrated superior performance compared to the baseline and other participating teams' scores in both the multi-lingual and English-translated test datasets, resulting in achieving the 1 𝑠𝑡 place in sub-task 1. Despite scoring lower than the baseline in sub-task 2 in the English test dataset, our submission for the multilingual test dataset surpassed the baseline score. Notably, the XLM-RoBERTaxl model, leveraging context and special tokens without class positive weights and fine-tuned on the combined training and validation data, exhibited strong performance in both sub-tasks for the multilingual data. Furthermore, our findings indicated that while class positive weights augmented the models' ability to classify under-represented classes, they did not yield an overall performance improvement. The shared task posed a significant challenge due to the presence of data imbalance across classes and languages, as well as the existence of low-resource languages in the texts.</p><p>To further optimize model performance for multi-label human value detection, future endeavors should center on exploring additional Transformer layers within the custom multi-head architecture, with a particular emphasis on even larger Transformer language models such as the XLM-RoBERTa-xxl 10 . Additionally, the investigation of alternative loss functions to address data imbalance, the implementation of data augmentation methods or even an ensemble of various models hold the potential to further enhance performance.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Limitations</head><p>The experimentation process in both sub-tasks has revealed a significant issue of class imbalance. Despite the assignment of higher weights to the minority classes, it has become evident that detecting one or more human values is a challenging task. This challenge primarily stems from the imbalance in the annotated human values across languages as well as the general class imbalance among human values in the multi-lingual training dataset. Moreover, the presence of low-resource languages such as Hebrew and Greek has posed a further challenge, as the multi-lingual models contain a smaller number of tokens for these languages in comparison to English. Notwithstanding these challenges, the multilingual models have performed adequately compared to the baseline models. Moreover, in the process of fine-tuning the XLM-RoBERTa-xl and DeBERTa-v2-xxl models, we encountered challenges stemming from limitations in GPU VRAM memory and time. Specifically, we modified the fine-tuning approach for the first model by reducing the number of Transformer layers from three to one. Furthermore, in the second model's case, the custom head's multi-head architecture did not incorporate any Transformer layers.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A. Appendix Table 4</head><p>Achieved F 1 -score of the baseline XLM-RoBERTa-base on the validation dataset for sub-task 1. The model was fine-tuned using the whole multi-lingual training dataset and evaluated in multi-lingual validation dataset.   </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Languages</head></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>1 HEFigure 2 : 1 to 𝑇 − 2 𝑘 and 𝑇 − 1 1to 𝑇 − 1</head><label>121211</label><figDesc>Figure 2: The figure presents the custom architecture that was used for the experiments. The model input consists of the [CLS] token, the tokens of the two preceding sentences (when available), (𝑇 −2 1 to 𝑇 −2 𝑘 and 𝑇 −1 1to 𝑇 −1 𝑘 , respectively) along with their corresponding special tokens ([𝑆𝑇−2 ] and [𝑆𝑇 −1 ], respectively), and the tokens of the sentence to be classified (𝑇 1 to 𝑇 𝑚 ). The [ST] token is a special token that corresponds to one of the 19 (or 38) classes. Each custom Transformer head corresponding to each language had three Transformer Layers for RoBERTa-large and the corresponding XLM variant, one Transformer layer for XLM-RoBERTa-xlarge, while DeBERTa-v2-xxl did not have any Transformer layers.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Radar Plot of 19 value categories (sub-task 1) illustrating the different performance per value in F1 score of our best-performing model compared to the baselines.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Number of texts and labels per language.</figDesc><table><row><cell></cell><cell cols="10">All English Greek German French Italian Dutch Bulgarian Turkish Hebrew</cell></row><row><cell>No. Texts</cell><cell>2648</cell><cell>408</cell><cell>328</cell><cell>261</cell><cell>219</cell><cell>276</cell><cell>323</cell><cell>260</cell><cell>323</cell><cell>250</cell></row><row><cell cols="2">No. Labels 27038</cell><cell>2590</cell><cell>3658</cell><cell>3876</cell><cell>1111</cell><cell>3262</cell><cell>3265</cell><cell>3624</cell><cell>4284</cell><cell>4992</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>Achieved F 1 -score of each submission on the test dataset for sub-task 1. A ✓ indicates that the submission used the automatic translation to English. Baseline submissions shown in gray.</figDesc><table><row><cell>F1-score</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>direction: thought Self-direction: action Stimulation Hedonism Achievement Power: dominance Power: resources Face Security: personal Security: societal Tradition Conformity: rules Conformity: interpersonal Humility Benevolence: caring Benevolence: dependability Universalism: concern Universalism: nature Universalism: tolerance</head><label></label><figDesc></figDesc><table><row><cell>multi-lingual XLM-RoBERTa-large_weights_context_ special tokens_19_only train data</cell><cell></cell><cell>34 13 20 28 28 37 37 45 22 33 46 46 49 21 04 32 32 47 63 21</cell></row><row><cell>multi-lingual XLM-RoBERTa-large_context_19</cell><cell></cell><cell>36 15 28 35 35 44 39 47 28 40 48 49 50 20 08 33 32 47 60 24</cell></row><row><cell>multi-lingual XLM-RoBERTa-xl_context_special tokens_19</cell><cell></cell><cell>38 15 27 31 36 43 41 51 32 44 49 48 51 23 00 34 35 50 63 24</cell></row><row><cell>multi-lingual XLM-RoBERTa-xl_context_special tokens_38</cell><cell></cell><cell>39 15 27 30 37 45 42 49 31 42 49 46 51 24 00 34 33 47 63 27</cell></row><row><cell>translated XLM-RoBERTa-large_context_special tokens_19</cell><cell>✓</cell><cell>35 14 25 30 28 41 40 46 25 40 48 48 48 20 05 34 30 46 59 25</cell></row><row><cell>translated RoBERTa-large_weights_context_special tokens_19_only train data</cell><cell>✓</cell><cell>37 19 23 31 32 40 41 45 31 43 48 51 48 26 11 34 33 48 60 27</cell></row><row><cell>translated RoBERTa-large_context_special tokens_19</cell><cell>✓</cell><cell>37 16 28 33 35 43 38 48 28 44 48 51 49 27 05 34 27 48 61 27</cell></row><row><cell>translated DeBERTa-v2-xxl_context_special tokens_19_only train data</cell><cell>✓</cell><cell>37 15 26 32 32 44 40 45 32 41 47 49 50 24 05 34 33 48 62 27</cell></row><row><cell>translated RoBERTa-large_context_special tokens_38</cell><cell>✓</cell><cell>37 12 24 32 36 42 39 46 28 43 47 49 49 22 00 34 32 47 61 27</cell></row><row><cell>valueeval24-bert-baseline-en</cell><cell>✓</cell><cell>24 00 13 24 16 32 27 35 08 24 40 46 42 00 00 18 22 37 55 02</cell></row><row><cell>valueeval24-random-baseline</cell><cell></cell><cell>06 02 07 05 02 11 08 10 04 05 13 03 11 03 00 04 04 09 04 02</cell></row><row><cell>valueeval24-random-baseline</cell><cell>✓</cell><cell>06 02 07 05 02 11 08 10 03 04 14 03 11 03 00 05 04 09 04 02</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 3</head><label>3</label><figDesc>Achieved F 1 -score of each submission on the test dataset for sub-task 2. A ✓ indicates that the submission used the automatic translation to English. Baseline submissions shown in gray.RoBERTa-large for the English dataset, both leveraging context and special tokens without class positive weights and having been fine-tuned on the combined training and validation data as the training set, resulted in identical Macro-F1 scores across all 38 classes (77%). Once again, the class with the lowest F1 score was Humility, scoring 25% and 22% in the multi-lingual and English test datasets, respectively, significantly lower than the baselines' scores. Conversely, the Universalism: nature value yielded the highest F1 scores in both of our submissions. Finally, the Universalism: tolerance value was once again successfully detected by the models, despite being underrepresented in the data, achieving 71% and 74% in the multi-lingual and English test datasets, respectively.</figDesc><table><row><cell>F 1 -score</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>All English Greek German French Italian Dutch Bulgarian Turkish Hebrew</head><label></label><figDesc></figDesc><table><row><cell>XLM-RoBERTa-base 29,5</cell><cell>22,41</cell><cell>26,16</cell><cell>25,24</cell><cell>2,52</cell><cell>22,71</cell><cell>18,71</cell><cell>23,30</cell><cell>28,03</cell><cell>24,16</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 5</head><label>5</label><figDesc>Models' Hyperparameters used for experiments and submissions. For the XLM-RoBERTa-xl model, 4 batch size was used for training, validation and testing as well as bf16 for mixed precision training. For the DeBERTa-v2-xxl model, 4 batch size was used for training, validation and testing as well as fp16 for mixed precision training.</figDesc><table><row><cell>Hyperparameter</cell><cell>Value</cell></row><row><cell>Seed</cell><cell>2024</cell></row><row><cell>Number of Epochs</cell><cell>20</cell></row><row><cell>Early Stopping Patience</cell><cell>5</cell></row><row><cell>Sequence Length</cell><cell>512</cell></row><row><cell>Train Batch Size</cell><cell>8 / 4</cell></row><row><cell>Validation / Test Batch Size</cell><cell>8 / 4</cell></row><row><cell>Learning Rate</cell><cell>5e-6</cell></row><row><cell>Weight Decay</cell><cell>0.01</cell></row><row><cell>Warm-up Ratio</cell><cell>0.01</cell></row><row><cell>Optimizer</cell><cell>AdamW</cell></row><row><cell>AdamW Epsilon</cell><cell>1e-8</cell></row><row><cell>LR Scheduler</cell><cell>Linear</cell></row><row><cell>Mixed Precision</cell><cell>fp16 / bf16</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_0">https://github.com/SotirisLegkas/Touche-ValueEval24-Hierocles-of-Alexandria</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_1">https://huggingface.co/FacebookAI/xlm-roberta-base</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_2">https://huggingface.co/FacebookAI/xlm-roberta-large</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_3">https://huggingface.co/facebook/xlm-roberta-xl</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_4">https://huggingface.co/FacebookAI/roberta-large</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_5">https://huggingface.co/microsoft/deberta-v2-xxlarge</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_6">https://pytorch.org/docs/stable/generated/torch.nn.functional.binary_cross_entropy_with_logits.html</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>The research leading to these results has received funding from the European Union's Horizon Europe research and innovation programme, in the context of: TITAN project, under grant agreement No. 101070658 and AI4TRUST project, under grant agreement No. 101070190. This paper reflects only the view of the authors and the European Commission is not responsible for any use that may be made of the information it contains.</p></div>
			</div>


			<div type="funding">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>1 https://www.vast-project.eu/ 2 https://knowledge4policy.ec.europa.eu/projects-activities/valuesml-unravelling-expressed-values-media-informed-policy-making_ en</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Refining the Theory of Basic Individual Values</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">H</forename><surname>Schwartz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Cieciuch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Vecchione</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Davidov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Fischer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Beierlein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ramos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Verkasalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J.-E</forename><surname>Lönnqvist</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Demirutku</surname></persName>
		</author>
		<idno type="DOI">10.1037/a0029393</idno>
	</analytic>
	<monogr>
		<title level="j">Journal of personality and social psychology</title>
		<imprint>
			<biblScope unit="volume">103</biblScope>
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">The four components of acting morally, Moral behavior and moral development</title>
		<author>
			<persName><forename type="first">D</forename><surname>Narvaez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Rest</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">An introduction</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="385" to="400" />
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">A knowledge graph of values across space and time</title>
		<author>
			<persName><forename type="first">M</forename><surname>Ruskov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Dagioglou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kokol</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Montanelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Petasis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">CEUR Workshop Proceedings</title>
				<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="volume">3536</biblScope>
			<biblScope unit="page" from="8" to="20" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Scharfbillig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Smillie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Mair</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sienkiewicz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Keimer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Pinho Dos Santos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Vinagreiro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Alves</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Vecchione</surname></persName>
		</author>
		<author>
			<persName><surname>Scheunemann</surname></persName>
		</author>
		<idno type="DOI">10.2760/349527</idno>
		<title level="m">Values and Identities -a Policymaker&apos;s Guide</title>
				<meeting><address><addrLine>Luxembourg</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
		<respStmt>
			<orgName>European Commission&apos;s Joint Research Centre</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical Report</note>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Attributions toward artificial agents in a modified moral turing test</title>
		<author>
			<persName><forename type="first">E</forename><surname>Aharoni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Fernandes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Brady</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Alexander</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Criner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Queen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Rando</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Nahmias</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Crespo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Scientific Reports</title>
		<imprint>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="page">8458</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Overview of Touché 2024: Argumentation Systems</title>
		<author>
			<persName><forename type="first">J</forename><surname>Kiesel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ç</forename><surname>Çöltekin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Heinrich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Alshomary</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">D</forename><surname>Longueville</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Erjavec</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Handke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kopp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ljubešić</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Meden</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Mirzakhmedova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Morkevičius</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Reitis-Münstermann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Scharfbillig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Stefanovitch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wachsmuth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<ptr target="//huggingface.co/facebook/xlm-roberta-xxlMeetsMultilinguality" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Fifteenth International Conference of the CLEF Association (CLEF 2024</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Mulhem</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Quénot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Schwab</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Soulier</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><forename type="middle">M D</forename><surname>Nunzio</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">G S</forename><surname>De Herrera</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<meeting>the Fifteenth International Conference of the CLEF Association (CLEF 2024<address><addrLine>Berlin Heidelberg New York</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note>Multimodality, and Interaction</note>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Andronicus of rhodes at SemEval-2023 task 4: Transformer-based human value detection using four different neural network architectures</title>
		<author>
			<persName><forename type="first">G</forename><surname>Papadopoulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kokol</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Dagioglou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Petasis</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2023.semeval-1.75</idno>
		<ptr target="https://aclanthology.org/2023.semeval-1.75.doi:10.18653/v1/2023.semeval-1.75" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 17th International Workshop on Semantic Evaluation (SemEval-2023), Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Ojha</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Doğruöz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Tayyar</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Madabushi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><surname>Kumar</surname></persName>
		</editor>
		<editor>
			<persName><surname>Sartori</surname></persName>
		</editor>
		<meeting>the 17th International Workshop on Semantic Evaluation (SemEval-2023), Association for Computational Linguistics<address><addrLine>Toronto, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="542" to="548" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">The ellogon web annotation tool: Annotating moral values and arguments</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">F</forename><surname>Ntogramatzis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gradou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Petasis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kokol</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Thirteenth Language Resources and Evaluation Conference</title>
				<meeting>the Thirteenth Language Resources and Evaluation Conference</meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="3442" to="3450" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">SemEval-2023 Task 4: ValueEval: Identification of Human Values behind Arguments</title>
		<author>
			<persName><forename type="first">J</forename><surname>Kiesel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Alshomary</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Mirzakhmedova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Heinrich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Handke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wachsmuth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2023.semeval-1.313</idno>
	</analytic>
	<monogr>
		<title level="m">17th International Workshop on Semantic Evaluation (SemEval 2023), Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">R</forename><surname>Kumar</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Ojha</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Doğruöz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><forename type="middle">D S</forename><surname>Martino</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><forename type="middle">T</forename><surname>Madabushi</surname></persName>
		</editor>
		<meeting><address><addrLine>Toronto, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="2287" to="2303" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Identifying the human values behind arguments</title>
		<author>
			<persName><forename type="first">J</forename><surname>Kiesel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Alshomary</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Handke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Cai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wachsmuth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2022.acl-long.306</idno>
		<ptr target="https://aclanthology.org/2022.acl-long.306.doi:10.18653/v1/2022.acl-long.306" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics</title>
		<title level="s">Long Papers</title>
		<editor>
			<persName><forename type="first">S</forename><surname>Muresan</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Villavicencio</surname></persName>
		</editor>
		<meeting>the 60th Annual Meeting of the Association for Computational Linguistics<address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="4459" to="4471" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Attention-based lstm, gru and cnn for short text classification</title>
		<author>
			<persName><forename type="first">S</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Zhao</surname></persName>
		</author>
		<idno type="DOI">10.3233/JIFS-191171</idno>
		<ptr target="https://doi.org/10.3233/JIFS-191171.doi:10.3233/JIFS-191171" />
	</analytic>
	<monogr>
		<title level="j">J. Intell. Fuzzy Syst</title>
		<imprint>
			<biblScope unit="volume">39</biblScope>
			<biblScope unit="page" from="333" to="340" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><surname>Cortiz</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2104.02041</idno>
		<title level="m">Exploring transformers in emotion recognition: a comparison of bert, distillbert, roberta, xlnet and electra</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">S</forename><surname>Brown</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Schneider</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">D</forename><surname>Dragan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Niekum</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2012.01557</idno>
		<title level="m">Value alignment verification</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">John-arthur at SemEval-2023 task 4: Fine-tuning large language models for arguments classification</title>
		<author>
			<persName><forename type="first">G</forename><surname>Balikas</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2023.semeval-1.197</idno>
		<ptr target="https://aclanthology.org/2023.semeval-1.197.doi:10.18653/v1/2023.semeval-1.197" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 17th International Workshop on Semantic Evaluation (SemEval-2023), Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Ojha</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Doğruöz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Tayyar</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Madabushi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><surname>Kumar</surname></persName>
		</editor>
		<editor>
			<persName><surname>Sartori</surname></persName>
		</editor>
		<meeting>the 17th International Workshop on Semantic Evaluation (SemEval-2023), Association for Computational Linguistics<address><addrLine>Toronto, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="1428" to="1432" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">PAI at SemEval-2023 task 4: A general multi-label classification system with class-balanced loss function and ensemble module</title>
		<author>
			<persName><forename type="first">L</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Li</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2023.semeval-1.34</idno>
		<ptr target="https://aclanthology.org/2023.semeval-1.34.doi:10.18653/v1/2023.semeval-1.34" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 17th International Workshop on Semantic Evaluation (SemEval-2023), Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Ojha</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Doğruöz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Tayyar</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Madabushi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><surname>Kumar</surname></persName>
		</editor>
		<editor>
			<persName><surname>Sartori</surname></persName>
		</editor>
		<meeting>the 17th International Workshop on Semantic Evaluation (SemEval-2023), Association for Computational Linguistics<address><addrLine>Toronto, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="256" to="261" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Adam-smith at SemEval-2023 task 4: Discovering human values in arguments with ensembles of transformer-based models</title>
		<author>
			<persName><forename type="first">D</forename><surname>Schroter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dementieva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Groh</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2023.semeval-1.74</idno>
		<ptr target="https://aclanthology.org/2023.semeval-1.74.doi:10.18653/v1/2023.semeval-1.74" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 17th International Workshop on Semantic Evaluation (SemEval-2023), Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Ojha</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Doğruöz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Tayyar</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Madabushi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><surname>Kumar</surname></persName>
		</editor>
		<editor>
			<persName><surname>Sartori</surname></persName>
		</editor>
		<meeting>the 17th International Workshop on Semantic Evaluation (SemEval-2023), Association for Computational Linguistics<address><addrLine>Toronto, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="532" to="541" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Mao-zedong at SemEval-2023 task 4: Label represention multihead attention model with contrastive learning-enhanced nearest neighbor mechanism for multilabel text classification</title>
		<author>
			<persName><forename type="first">C</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Xiao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Fei</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2023.semeval-1.58</idno>
		<ptr target="https://aclanthology.org/2023.semeval-1.58.doi:10.18653/v1/2023.semeval-1.58" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 17th International Workshop on Semantic Evaluation (SemEval-2023), Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Ojha</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Doğruöz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Tayyar</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Madabushi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><surname>Kumar</surname></persName>
		</editor>
		<editor>
			<persName><surname>Sartori</surname></persName>
		</editor>
		<meeting>the 17th International Workshop on Semantic Evaluation (SemEval-2023), Association for Computational Linguistics<address><addrLine>Toronto, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="426" to="432" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<title level="m" type="main">Attention is all you need</title>
		<author>
			<persName><forename type="first">A</forename><surname>Vaswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shazeer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Parmar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Uszkoreit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">N</forename><surname>Gomez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Kaiser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Polosukhin</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1706.03762</idno>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Continuous Integration for Reproducible Shared Tasks with TIRA</title>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wiegmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Kolyada</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Grahm</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Elstner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Loebe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-031-28241-6_20</idno>
	</analytic>
	<monogr>
		<title level="m">Advances in Information Retrieval. 45th European Conference on IR Research (ECIR 2023)</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Crestani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Maistro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Joho</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Davis</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Gurrin</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">U</forename><surname>Kruschwitz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Caputo</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin Heidelberg New York</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="236" to="241" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Identifying the Human Values behind Arguments</title>
		<author>
			<persName><forename type="first">J</forename><surname>Kiesel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Alshomary</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Handke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Cai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wachsmuth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2022.acl-long.306</idno>
	</analytic>
	<monogr>
		<title level="m">60th Annual Meeting of the Association for Computational Linguistics (ACL 2022), Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">S</forename><surname>Muresan</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Villavicencio</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="4459" to="4471" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<title level="m" type="main">The touché23-valueeval dataset for identifying human values behind arguments</title>
		<author>
			<persName><forename type="first">N</forename><surname>Mirzakhmedova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kiesel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Alshomary</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Heinrich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Handke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Cai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Valentin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dastgheib</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Ghahroodi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Sadraei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Asgari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Kawaletz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wachsmuth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2301.13771.arXiv:2301.13771" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<title level="m" type="main">Roberta: A robustly optimized bert pretraining approach</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ott</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Joshi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Levy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Stoyanov</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1907.11692</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Conneau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Khandelwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Chaudhary</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Wenzek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Guzmán</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Grave</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ott</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Stoyanov</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1911.02116</idno>
		<title level="m">Unsupervised cross-lingual representation learning at scale</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<monogr>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ott</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Anantharaman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Conneau</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2105.00572</idno>
		<title level="m">Larger-scale transformers for multilingual masked language modeling</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Chen</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2006.03654</idno>
		<title level="m">Deberta: Decoding-enhanced bert with disentangled attention</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<monogr>
		<title level="m" type="main">Focal loss for dense object detection</title>
		<author>
			<persName><forename type="first">T.-Y</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Girshick</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Dollár</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/1708.02002.arXiv:1708.02002" />
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<monogr>
		<title level="m" type="main">Class-balanced loss based on effective number of samples</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Cui</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Jia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T.-Y</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Belongie</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/1901.05555.arXiv:1901.05555" />
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<monogr>
		<title level="m" type="main">Distribution-balanced loss for multi-label classification in long-tailed datasets</title>
		<author>
			<persName><forename type="first">T</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Lin</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2007.09654.arXiv:2007.09654" />
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<monogr>
		<title level="m" type="main">Balancing methods for multi-label text classification with long-tailed class distribution</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Giledereli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Köksal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Özgür</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Ozkirimli</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2109.04712" />
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
