<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Team INSA Passau at Touché: Multi-lingual Parliamentary Speech Classification Notebook for the Touché Lab at CLEF 2024</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Maud</forename><surname>Andruszak</surname></persName>
							<email>maud.andruszak@insa-lyon.fr</email>
							<affiliation key="aff0">
								<orgName type="institution">INSA de Lyon</orgName>
								<address>
									<addrLine>20 Avenue Albert Einstein</addrLine>
									<postCode>69100</postCode>
									<settlement>Villeurbanne</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">Universität Passau</orgName>
								<address>
									<addrLine>Innstraße 41</addrLine>
									<postCode>94032</postCode>
									<settlement>Passau</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Alaa</forename><surname>Alhamzeh</surname></persName>
							<email>alaa.alhamzeh@uni-passau.de</email>
							<affiliation key="aff1">
								<orgName type="institution">Universität Passau</orgName>
								<address>
									<addrLine>Innstraße 41</addrLine>
									<postCode>94032</postCode>
									<settlement>Passau</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Előd</forename><surname>Egyed-Zsigmond</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">INSA de Lyon</orgName>
								<address>
									<addrLine>20 Avenue Albert Einstein</addrLine>
									<postCode>69100</postCode>
									<settlement>Villeurbanne</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Anton</forename><surname>Carlsson</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">INSA de Lyon</orgName>
								<address>
									<addrLine>20 Avenue Albert Einstein</addrLine>
									<postCode>69100</postCode>
									<settlement>Villeurbanne</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Johan</forename><surname>Leydet</surname></persName>
							<email>johan.leydet@insa-lyon.fr</email>
							<affiliation key="aff0">
								<orgName type="institution">INSA de Lyon</orgName>
								<address>
									<addrLine>20 Avenue Albert Einstein</addrLine>
									<postCode>69100</postCode>
									<settlement>Villeurbanne</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Yasser</forename><surname>Otiefy</surname></persName>
							<email>yasser.otiefy@uni-passau.de</email>
							<affiliation key="aff1">
								<orgName type="institution">Universität Passau</orgName>
								<address>
									<addrLine>Innstraße 41</addrLine>
									<postCode>94032</postCode>
									<settlement>Passau</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff2">
								<address>
									<settlement>Grenoble</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Team INSA Passau at Touché: Multi-lingual Parliamentary Speech Classification Notebook for the Touché Lab at CLEF 2024</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">67E69A6E9BA78837AC1D495E2CC5CB70</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:01+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Political debates</term>
					<term>Large language models (LLMs)</term>
					<term>Few shot learning</term>
					<term>Text classification</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In this paper, we present the architecture used for our participation in the shared task of Ideology and Power Identification in Parliamentary Debates -Touché Lab at CLEF 2024. This task aims to identify the ideology of the speaker's party, and to identify whether the speaker's party is currently governing or in opposition. Furthermore, the data associated with these two sub-tasks are proposed from a multilingual perspective, where the speeches belong to at least 29 national or regional parliaments. Among our submitted runs, we achieved the best performance through BERT fine-tuning for both sub-tasks, in addition to Llama 3 prompting for a subset of the parliaments on identifying the speaker's party.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The identification of ideology and power structures in parliamentary debates is of vital importance for a comprehensive understanding of political dynamics and decision-making processes. Recognizing the background of ideological stances and power relations within debates facilitates a deeper insight into the strategic maneuvers of policymakers and the potential biases influencing legislative outcomes. Given the extensive volume of debates and the complexity of political discourse, the need to automate this identification process is increasingly urgent. Automation not only enhances the efficiency of analysis, but it may also ensure a more objective and consistent evaluation of the data, devoid of human biases.</p><p>The current edition of the Touché shared task <ref type="bibr" target="#b0">[1]</ref> tackles these issues by suggesting two sub-tasks: Sub-Task 1: Given a parliamentary speech in one of several languages, identify the ideology of the speaker's party. Sub-Task 2: Given a parliamentary speech in one of several languages, identify whether the speaker's party is currently governing or in opposition.</p><p>We have participated in both sub-tasks as team INSA.</p><p>The baseline defined by the organizers of the task is a linear logistic regression, using the term frequency-inverse document frequency (TF-IDF) to process the texts. The results using this baseline differ a lot between the parliaments, but also from the same parliament between the two sub-tasks. This entails that some approaches may perform better for some parliaments/sub-tasks than another. In addition, <ref type="bibr">Alhamzeh et al.</ref> showed in an intensive empirical study <ref type="bibr" target="#b1">[2]</ref> on text classification tasks, that some classical machine learning methods (specifically SVM) can outperform more complex deep learning models (BERT in their case). Hence, we decided to study varied methods.</p><p>Briefly, we have implemented four approaches. Two of them are based on Large Language Models (LLMs): the first one consists of fine-tuning a LLM, which we did with BERT as well as Llama 3, and the second one is based on prompting an LLM. Our third approach is based on manual features and is linked to the frequency of discriminatory words in a text, and the last one is a Support Vector Machine (SVM).</p><p>This paper is organized as follows: All of our submitted approaches will be presented in Section 2. We elaborate on the outcomes for both subtasks and different parliaments in Section 3. Finally, we conclude our work in Section 4.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Methods</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Fine --tuning pre-trained LLM</head><p>Our first approach consists of fine tuning a Large Language Model (LLM), which is an advanced artificial intelligence system designed to understand and generate human language. Utilizing deep learning and neural network architectures like transformers, these models are trained on extensive text datasets, allowing them to predict and generate coherent text based on context. LLMs excel in various tasks such as text generation, translation, summarizing, question answering, and sentiment analysis, making them invaluable for applications like chatbots, content creation, education, and medical diagnosis. They represent a remarkable leap in enabling sophisticated human-machine language interactions.</p><p>To train and test our models, we split the train dataset in two parts: 80% for training and 20% for testing. To match the task evaluation strategy, we evaluated our runs using the macro-averaged F1 score, which calculates the harmonic mean of precision and recall while assigning equal weight to all classes.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.1.">BERT</head><p>Among available open source LLMs, we examined the Bidirectional Encoder Representations from Transformers (BERT) <ref type="bibr" target="#b2">[3]</ref>. BERT is a transformer that can be fine-tuned by adding an output layer. On Hugging-Face, <ref type="foot" target="#foot_0">1</ref> we can find a lot of transformers based on BERT. We looked at several of them, trying to find the one that fitted our task the best. We studied the following transformers, some of them general like bert-base-cased, bert-base-uncased, roberta-base, and some others trained on law data like nlpaueb/legal-bert-base-uncased <ref type="bibr" target="#b3">[4]</ref>, casehold/legalbert <ref type="bibr" target="#b4">[5]</ref>, saibo/legal-roberta-base<ref type="foot" target="#foot_1">2</ref> or pile-oflaw/legalbert-large-1.7M-2 <ref type="bibr" target="#b5">[6]</ref>. All of these are trained to support English written texts. Consequently, we tested those transformers on data from the British parliament (GB). Our system for this approach consists of five steps, as shown in Figure <ref type="figure" target="#fig_0">1</ref>. First, we use the tokenizer corresponding to our LLM transformer to tokenize the input texts. Then, our BERT classifier takes these new inputs to be applied on the LLM transformer. After this, the transformer output goes through a dropout layer, a linear transformation with an output of dimension 1 and finally a sigmoid layer. The dropout layer consists of cutting nodes from the input and hidden layers of the neural network, which avoids overfitting <ref type="bibr" target="#b6">[7]</ref>. The sigmoid layer maps the input to a floating point between 0 and 1, which is easily suitable for binary classification <ref type="bibr" target="#b7">[8]</ref>. This floating point is our final result. The hard label is computed with a simple threshold set at 0.5, as it is on the task baseline.</p><p>The experiment was carried out with 3 folds cross-validation. The parameters searched for optimization were the LLM transformer as well as the dropout probability, the number of epochs, the learning rate and the tokens length. We ended up taking the bert-base-uncased transformer, which showed the best F1 score on this experiment. Our dropout probability was set to 0.2, the number of epochs to 3, the learning rate to 0.00001, and the tokens length to 150. As for the second step, we optimized the system by choosing the best loss function. We tested the cross-entropy loss and the binary cross-entropy loss. The latter was the best. With all these parameters, we created a baseline with BERT for every parliament, by training each of them separately on their English translations.</p><p>With our BERT baseline results, we noticed that not all parliament texts behave the same when put inside a BERT model. We have improvements of the F1 score for some countries, but not for all of them. Then, we focused on improving the results by augmenting the training data. To do so, we firstly trained a single model on the English translations of every parliament at the same time. After this, we tried to increase the F1 score while having the minimal best training data, meaning that we may not need to add data from all parliaments to have better results. We did this with two things in mind : energy consumption, as well as time efficiency. Indeed, less training data decreases the time needed to train the model, and some parliaments might not help in a certain parliament predictions. That's why we implemented an incremental process to add a parliament's data to the training data until we decide that adding some data is useless, starting from the sole studied parliament's data. For this, we trained every pair of parliaments of a same sub-task. Then, for each parliament, we sorted the pairs containing this parliament in descending order of F1 score. After that, we used this order to incrementally add a parliament to the training, each time adding the one from the pair with the best F1 score decrementing. As explained above, we stopped the incremental process when it was not helping anymore.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.2.">Llama 3</head><p>Llama 3 <ref type="bibr" target="#b8">[9]</ref> is the latest model of Meta's LLM series so far/to date. It significantly enhances the capabilities introduced by its predecessors. This model was released on April 18, 2024 and comes in multiple configurations, including 8 billion and 70 billion parameters, optimized for both performance and scalability. Notably, Llama 3 has been trained on an extensive dataset of up to 15 trillion tokens, demonstrating improved performance even with smaller, more efficient models. This efficiency allows it to generate high-quality results comparable to larger models while being easier to deploy and run on standard hardware configurations.</p><p>We only tested Llama 3 on the GB parliament data of sub-task 1 (orientation). We chose this parliament because the original language of this dataset is English, and Llama 3 is trained mainly on English-speaking data. The hyperparameter tuning step was realized with a grid search on a 5 folds cross-validation and resulted with this configuration:</p><p>• Model name: "meta-llama/Meta-Llama-3-8B" • Number of epochs: 5 • Learning rate: 0.00005 • Maximum tokens length: 256 This configuration yielded a mean F1 score of 0.686.</p><p>We could not find enough time to train and test on other parliaments. Even if the mean F1 score is not really high, we decided to submit this run with the best configuration of this approach.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Prompting</head><p>Prompt engineering is the process of crafting and refining the inputs given to a large language model (LLM) to achieve desired outputs. This technique involves carefully designing the phrasing, structure, and context of prompts to guide the model's responses effectively. By manipulating prompts, users can optimize the performance of LLMs in various applications, such as generating specific types of content, answering questions accurately, or performing complex tasks. Prompt engineering requires a deep understanding of how LLMs interpret and process language, enabling users to harness the full potential of these models while minimizing errors and biases in their outputs. This practice is crucial for maximizing the utility and accuracy of LLMs in real-world scenarios.</p><p>For the sub-task 1 (orientation), we are employing Llama 3, the latest generative text LLM made by Meta. The approach focuses on prompt engineering to guide the model to the specific task at hand.</p><p>The base prompt is composed of three distinct parts: the definition of the task, the sentence that we want to classify and the expected labels.</p><p>Here are the specific strategies we implemented: • Voting approach: We combine the results from multiple prompt variations to make a final classification decision based on the majority vote, thereby improving the robustness of the predictions. In this voting method, we use three predictions: the zero-shot learning prediction and two different few-shot learning prediction with mixed correctness. • Zero-shot learning with labels explained: We enhance the zero-shot prompts by including an explanatory sentence about the political labels. This sentence clarifies what "left" and "right" typically refer to, as follows:</p><p>"Usually, Left advocates for social equality through government intervention and prioritizes issues like economic redistribution and social justice, while Right emphasizes individual liberty, free market principles, and traditional values, often favoring limited government intervention and policies that promote economic freedom. "</p><p>As we have aforementioned, there are two variants of Llama 3 model: either with 8 billion parameters (8b) or with 70 billion parameters (70b). We have tested the zero-shot learning approach on both of them, and immediately realized the result differences between them. We tested those on 20% of the GB orientation dataset, that is to say 4 848 rows. We respectively obtained F1 scores of 0.758 and 0.811 with 8b and 70b. By testing these strategies on part of our training data, we determined the ones that could be useful. Thereby, we eliminated the few-shot learning strategy using incorrect predictions, as it was influencing the answer too much given the fact that every example was going in the opposite direction as what the model wants to answer. We also eliminated the few-shot learning strategy using the highest cosine similarity examples as the results were significantly worse than the zero-shot learning strategy. It seems that the examples with high similarity are not helping the model, we may just interfere in the prediction because of their length : these examples are 4 to 8 times longer than the average (2540 characters). This means they are not more discriminatory, only longer so they contain more words and thus have more similarity to the other texts.</p><p>The results of all our strategies kept for this approach, evaluated with the 70 billion parameters configuration <ref type="foot" target="#foot_2">3</ref> , can be found in Table <ref type="table" target="#tab_1">1</ref>. To balance the two different runs with the few-shot learning strategies, we put an example correctly predicted 'Left' and one labelled 'Right' but wrongly predicted 'Left' on the first one, and the other way around for the second run. It seems also important to notice that we can achieve decent results using this prompting method on the sub-task 1 (orientation), but it appeared to be much more complicated to gain anything from this approach on the sub-task 2 (power).</p><p>By employing these prompt engineering techniques, we aimed to leverage Llama 3's generative capabilities to improve the accuracy and reliability of political orientation classification. The careful design and testing of various prompt strategies allowed us to explore different aspects of the model's understanding and adaptability to the task at hand.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Manual Features</head><p>Besides LLMs, we have examined a more fundamental approach to the problem that is based on finding manual features, certain basic features of the texts that might be useful when classifying. Each manual feature results in one value for each text, which can then be used either as a complement to the prediction of the LLM, or as a prediction on its own.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.1.">Z-score</head><p>If given a corpus with two distinct parts, 𝑃 0 and 𝑃 1 , it is possible to calculate how over-and underused each lexical token in 𝑃 0 is in comparison to 𝑃 1 . The formula used is as follows:</p><formula xml:id="formula_0">Z − score(𝑡 𝑖𝑗 ) = 𝑡𝑓 𝑖𝑗 − 𝑛 𝑗 • 𝑝(𝑡 𝑖 ) √︀ 𝑛 𝑗 • 𝑝(𝑡 𝑖 ) • (1 − 𝑝(𝑡 𝑖 ))</formula><p>An explanation of the terms can be found in Table <ref type="table" target="#tab_2">2</ref>. A high z-score for token 𝑡 𝑖0 indicates that token 𝑖 is used more frequently than expected in corpus 𝑃 0 compared to corpus 𝑃 1 , and a low z-score indicates that the token is used less frequently in corpus 𝑃 0 than expected <ref type="bibr" target="#b9">[10]</ref>. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.2.">Running Z-score sum</head><p>Given that the data for both the orientation and the power task is divided into two separate labels, 0 and 1, it is possible to calculate the most over-and underused tokens for each label for each country. This was done from the point of view of label 0, meaning that a high z-score signifies that a token is overused in texts with label 0, and a low z-score signifies that a token is underused in texts with label 0.</p><p>For the corpus of each country, the z-score for each token was calculated. All tokens where the absolute amount of the z-score was above 2 were deemed discriminatory and thus usable to identify the label of a given text, but at most 5% of the total tokens for a country's corpus were allowed to be classified as discriminatory. Different discriminatory words were used for the orientation and power task for the same country.</p><p>As an example, we present in Table <ref type="table">3</ref> a few of the most discriminatory words from the Swedish dataset. Interestingly, they indicate that topics such as immigration, law and order, and economics are more frequently discussed by the right-wing parties, as indicated by the overuse by words such as migration, immigration, finance, police, and entrepreneurs. The left-wing parties on the other hand seem to focus their debates on equality, climate, and social benefits, since the overused words include climate, welfare, women, sustainable, and racism. Another interesting note is that bourgeois is included in the words overused by the left-wing parties. The term bourgeois has traditionally been used by the left in Sweden to address the right.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 3</head><p>A few of the most discriminatory words for the Swedish dataset. The two leftmost columns have words with z-scores less than zero, which are overused in datapoints with label 1. The two rightmost columns have z-scores greater than zero, which are overused in datapoints with label 0. For each text, two manual features were calculated using the discriminatory words:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Word</head><p>1. For each token that was considered discriminatory, its number of occurrences were added to a running sum if the z-score of the word was greater than zero, or subtracted from the running sum if the z-score was lower than zero.</p><p>2. For each token that was considered discriminatory, its z-score was multiplied by the number of occurrences, and that value was added to a running sum.</p><p>For both metrics, all the values were normalized to have a mean of zero and a standard deviation of 1. This was done country-wise.</p><p>To classify a given text, the optimal decision boundary was calculated. This was done by sliding the decision boundary from the lowest recorded running sum up to the highest recorded running sum, and for each decision boundary classifying all texts with a lower running sum as 1 and with a higher running sum as 0. The optimal decision boundary was considered the one that resulted in the highest F1 score. The datasets for some countries were very uneven, with upwards of 80% of the texts belonging to one label. To make sure that the optimal decision boundary was not just classifying everything as the majority label, for each country that had a label split of more than 60/40, texts from the majority label were randomly removed until a 50/50 split was achieved.</p><p>Since there were two different ways of calculating the running sum, for each language the one that resulted in the best F1 score was kept. Thus, for a given text from a given country, a prediction was made as follows:</p><p>1. Calculate the running z-score sum according to the best method for the country 2. Normalize the sum with the mean and standard deviation calculated earlier 3. If the normalized running sum is greater than the decision boundary for the given country, classify the text as label 0. Otherwise, classify the text as label 1.</p><p>Table <ref type="table">4</ref> displays the F1 scores achieved when using only the z-score sum with the optimal decision boundary to classify data. This was all done on perfectly balanced data sets. Figure <ref type="figure" target="#fig_1">2</ref> shows the distribution of the z-score sum for two different languages. As is visible, the mean of the different labels vary in both examples. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.3.">Combining with BERT</head><p>As we have several approaches that perform well on certain texts/parliaments, we wanted to take the best out of the different models. Therefore, for each parliament, we ran the BERT model as well as the model using the discriminatory words. Each BERT prediction is a floating point between 0 and 1. This smoothed result gives us a level of confidence of the model in its prediction. The closer to an extremum a prediction is, the more confident the model is. When using the discriminatory words model, we also</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 4</head><p>Results on the training data when using only the manual feature running Z-score sum with the optimal decision boundary, i.e. classifying all texts with a running Z-score sum less than the decision boundary as 1, and texts with a running Z-score sum greater than the decision boundary as 0. Classification was done on perfectly balanced data sets, meaning that the majority label has been randomly downsampled until a perfect 50/50 split between labels was achieved. Empty cells are parliaments absent from a sub-task. get a confidence value. To aggregate those two results, we decided to select the prediction with the highest confidence.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Country F1-score on</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4.">SVM</head><p>Finally, we trained a Support Vector Machine (SVM). To explain briefly, SVM is a supervised learning algorithm used for classification and regression. It works by finding the hyperplane that best separates data into classes, maximizing the margin between the classes' closest points, called support vectors. SVMs can handle linear and non-linear data using kernel functions to transform the data into higher dimensions.</p><p>In our system, we preprocess the texts with a term frequency-inverse document frequency (TF-IDF) vectorizer. Then, we use these vectors in the SVM. The feature used is made of character n-grams, for which we chose to restrain the lower and upper boundaries of the n-grams to one and three respectively. This means that our feature is composed of unigrams, bigrams and trigrams. We decided to test this binary text classification with a fundamental approach, to compare with more complicated and elaborate approaches. In that sense, we did not try to boost this method's performances with a hyperparameter search, but just to have an idea of the power of this approach.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Results</head><p>In this section, we present our results for our different approaches. In Tables <ref type="table" target="#tab_7">5, 6 and 7</ref>, we use the following abbreviations for the approach name:</p><p>• Baseline: Baseline submitted by the organizers of the task • Bert-basic : Bert trained on the sole parliament's data • Bert-all-lang: Bert trained on every parliament's data • Bertaugm: Bert trained with the training's data incremental process • Llama3: Llama3-70b prompting with zero-shot learning • Svm: Support Vector Machine • Logreg: Logistic regression, which is our run on the baseline approach As we have aforementioned, we have not submitted all methods on all parliaments. Instead, prompting methods were used only on the GB orientation dataset, as well as the Llama 3 fine-tuning method. The logistic regression, SVM, and BERT fine-tuned methods were submitted to every parliament on both sub-tasks (or almost every parliament, some are missing because of mistakes). Finally, the BERT augmented approach has only been submitted on 10 parliaments on the power task: BA, DK, ES, ES-PV, GB, GR, HR, RS, SI, TR, as these were the ones showing a significant increase of the F1-score on our test data. Table <ref type="table" target="#tab_5">5</ref> exhibits an overview, for each one of our methods, the number of parliaments (on both subtasks taken together) where this precise method gave the best F1 score among all methods submitted. For example, out of the 53 parliaments, 14 of them had the best F1-score using SVM as a classifier.</p><p>Moreover, Table <ref type="table" target="#tab_6">6</ref> and Table <ref type="table" target="#tab_7">7</ref> detail our best submission outcomes for each parliament on the orientation and power sub-tasks, respectively. For each parliament, we can learn our best achieved F1-score along with the approach used to get this result. Additionally, we compare our result with the baseline, and report the improvement over it in terms of F1-score. For example, in Table <ref type="table" target="#tab_7">7</ref>, we can see that our best F1-score for the parliament HU on the power task is 0.89968, obtained using the SVM method. Compared to the baseline (F1-score of 0.857642), we enhanced this score by 3.2326%.</p><p>Evaluating the average F1-score on all parliaments of the sub-task 1 (orientation), our best method is BERT trained on all orientation parliaments. The submission of this strategy led to an F1-score of 0.585136, outperforming the baseline by 2.4829% and ranking our team eight out of ten. On the sub-task 2, our best method is the method of BERT with augmented training data, which achieved an F1-score of 0.625254, being 0.14802 behind the baseline, ranking us ten out of eleven.</p><p>Even though those results are not showing great improvement on the overall parliaments, some individual results are notable and show promising progress. For example, the SVM approach is also good for some parliaments. Overall, the average F1 score decreased compared to the baseline (around 1.5% for orientation and 3.5% for power), but it is important to note that it allowed us to increase the F1 score a lot on some parliaments, up to 17.6953% on the LV power parliament, as can be seen in Table <ref type="table" target="#tab_7">7</ref>.</p><p>The SVM method achieved great results on some parliaments of sub-task 2 (power): it led us to an F1 score of 0.889968 for the HU parliament, setting our team at the second place (out of ten) of the ranking; and achieved F1 scores of 0.880689 for ES-GA and 0.846702 for ES-CT, both setting us at the third place for these two parliaments. It is the same for sub-task 1 (orientation), where we achieved 0.692944 with the SVM, setting us to the second place (out of eleven). Fine-tuning BERT was found to be a beneficial approach for certain parliaments. The training of BERT using all available parliament data resulted in our team achieving the second position in the sub-task 1 for the BA parliament, achieving an F1 score of 0.526129. Significant results were also observed for a specific parliament in sub-task 1 (orientation). Prompting Llama 3 demonstrated effectiveness, particularly for the GB orientation parliament, achieving an F1 score of 0.790132. This approach secured the second position out of ten methods in the final ranking for this parliament. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Conclusion</head><p>In this paper, we described our different approaches for the ideology and power identification in parliamentary debates shared task, proposed by the Touché Lab at CLEF 2024.</p><p>We have reported several approaches, including fine-tuning different LLMs, prompting Llama 3, as well as examining a classical SVM which proved to be worth of interest. We discovered that one approach is not always working the best for every parliament, but can show interesting results for some of them. Because of this, we could achieve better results by creating a rule based approach, and using the method that individually works best for each parliament. We plan in our future work to integrate an argument mining phase in the classification pipeline, since it proved to be essential in different applications like comparative question answering <ref type="bibr" target="#b10">[11]</ref>, and financial speech analysis <ref type="bibr" target="#b11">[12]</ref>. Thus, we can use argumentation to analyze the given debate speech.</p><p>We also plan to extend on the prompting approach to non-English-speaking texts. Indeed, Llama 3 currently works with English-speaking data, but as Meta is working on training Llama 3 on multilingual data in a near future, it could become an effective solution for other languages soon. In addition, examining GPT capabilities on political speech analysis would be interesting.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Overview of our BERT fine tuning model</figDesc><graphic coords="2,139.69,504.55,315.89,200.22" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Boxplots of the z-score sum for two different countries.</figDesc><graphic coords="7,72.00,439.31,216.61,173.82" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1</head><label>1</label><figDesc>Llama-3-70b prompting results on sub-task 1 from British training data (GB)</figDesc><table><row><cell>Strategy</cell><cell cols="3">F1 score Precision Recall</cell></row><row><cell>zero-shot learning</cell><cell>0.811</cell><cell>0.83</cell><cell>0.814</cell></row><row><cell>few-shot learning with mixed correctness 1</cell><cell>0.791</cell><cell>0.813</cell><cell>0.794</cell></row><row><cell>few-shot learning with mixed correctness 2</cell><cell>0.799</cell><cell>0.814</cell><cell>0.802</cell></row><row><cell>voting</cell><cell>0.805</cell><cell>0.824</cell><cell>0.808</cell></row><row><cell>zero-shot learning with labels explained</cell><cell>0.776</cell><cell>0.807</cell><cell>0.781</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2</head><label>2</label><figDesc>Explanation of the terms used when calculating the Z-score. Index 𝑗 represents the corpus, and can have value either 0 or 1. Index 𝑖 represents the index of the current token, and can take values from 1 to 𝑁 , with 𝑁 being the number of tokens in the corpus.</figDesc><table><row><cell cols="2">Term Explanation</cell></row><row><cell>𝑡 𝑖𝑗</cell><cell>Token 𝑖, present in corpus 𝑗</cell></row><row><cell>𝑡𝑓 𝑖𝑗</cell><cell>Occurrence frequency of token 𝑖 in corpus 𝑗</cell></row><row><cell>𝑛 𝑗</cell><cell>Total number of tokens in corpus 𝑗</cell></row><row><cell>𝑝(𝑡 𝑖 )</cell><cell>Probability of token 𝑖 being selected when randomly sampling both corpora, estimated as (𝑡𝑓 𝑖0 + 𝑡𝑓 𝑖1 )/𝑛</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Subtask 1 F1-score on Subtask 2</head><label></label><figDesc></figDesc><table><row><cell>AT</cell><cell>0.6903</cell><cell>0.7071</cell></row><row><cell>BA</cell><cell>0.6960</cell><cell>0.6700</cell></row><row><cell>BE</cell><cell>0.7019</cell><cell>0.6757</cell></row><row><cell>BG</cell><cell>0.6668</cell><cell>0.7241</cell></row><row><cell>CZ</cell><cell>0.6750</cell><cell>0.6690</cell></row><row><cell>DK</cell><cell>0.6783</cell><cell>0.7085</cell></row><row><cell>EE</cell><cell>0.6971</cell><cell>-</cell></row><row><cell>ES</cell><cell>0.6970</cell><cell>0.6994</cell></row><row><cell>ES-CT</cell><cell>0.6992</cell><cell>0.7236</cell></row><row><cell>ES-GA</cell><cell>0.7712</cell><cell>0.7916</cell></row><row><cell>ES-PV</cell><cell>-</cell><cell>0.7335</cell></row><row><cell>FR</cell><cell>0.6901</cell><cell>0.6733</cell></row><row><cell>GB</cell><cell>0.7222</cell><cell>0.7334</cell></row><row><cell>GR</cell><cell>0.6958</cell><cell>0.6926</cell></row><row><cell>HR</cell><cell>0.6762</cell><cell>0.6654</cell></row><row><cell>HU</cell><cell>0.7355</cell><cell>0.7818</cell></row><row><cell>IS</cell><cell>0.7101</cell><cell>-</cell></row><row><cell>IT</cell><cell>0.6792</cell><cell>0.7012</cell></row><row><cell>LV</cell><cell>0.6834</cell><cell>0.6811</cell></row><row><cell>NL</cell><cell>0.6701</cell><cell>0.7108</cell></row><row><cell>NO</cell><cell>0.6704</cell><cell>-</cell></row><row><cell>PL</cell><cell>0.6888</cell><cell>0.6903</cell></row><row><cell>PT</cell><cell>0.6994</cell><cell>0.6802</cell></row><row><cell>RS</cell><cell>0.6763</cell><cell>0.7179</cell></row><row><cell>SE</cell><cell>0.7133</cell><cell>-</cell></row><row><cell>SI</cell><cell>0.6707</cell><cell>0.6835</cell></row><row><cell>TR</cell><cell>0.7053</cell><cell>0.7181</cell></row><row><cell>UA</cell><cell>0.7286</cell><cell>0.6931</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 5</head><label>5</label><figDesc>Number of parliaments where a certain approach is the best of all our approaches, both sub-tasks taken together</figDesc><table><row><cell>Approach</cell><cell>Parliaments Count</cell></row><row><cell>svm</cell><cell>14</cell></row><row><cell>bert-all-lang</cell><cell>11</cell></row><row><cell>baseline</cell><cell>10</cell></row><row><cell>logreg</cell><cell>8</cell></row><row><cell>bert-basic</cell><cell>6</cell></row><row><cell>bertaugm</cell><cell>3</cell></row><row><cell>Llama3</cell><cell>1</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 6</head><label>6</label><figDesc>Best Orientation Evaluation per Parliament</figDesc><table><row><cell cols="4">Parliament Our approach F1 Score F1 Difference to baseline</cell></row><row><cell>TR</cell><cell>baseline</cell><cell>0.840882</cell><cell>+0.0000%</cell></row><row><cell>GB</cell><cell>Llama3</cell><cell>0.790132</cell><cell>+4.5201%</cell></row><row><cell>ES-GA</cell><cell>svm</cell><cell>0.780211</cell><cell>+11.3434%</cell></row><row><cell>SE</cell><cell>baseline</cell><cell>0.749723</cell><cell>+0.0000%</cell></row><row><cell>GR</cell><cell>baseline</cell><cell>0.741625</cell><cell>+0.0000%</cell></row><row><cell>HU</cell><cell>svm</cell><cell>0.737012</cell><cell>+16.6735%</cell></row><row><cell>ES</cell><cell>logreg</cell><cell>0.71786</cell><cell>+0.0365%</cell></row><row><cell>ES-CT</cell><cell>svm</cell><cell>0.692944</cell><cell>+3.8107%</cell></row><row><cell>PL</cell><cell>bert-all-lang</cell><cell>0.692107</cell><cell>+23.2108%</cell></row><row><cell>UA</cell><cell>svm</cell><cell>0.682056</cell><cell>+9.7984%</cell></row><row><cell>RS</cell><cell>bert-basic</cell><cell>0.639734</cell><cell>+11.3972%</cell></row><row><cell>PT</cell><cell>logreg</cell><cell>0.63335</cell><cell>+0.1516%</cell></row><row><cell>NO</cell><cell>baseline</cell><cell>0.615736</cell><cell>+0.0000%</cell></row><row><cell>BG</cell><cell>bert-all-lang</cell><cell>0.613049</cell><cell>+7.9860%</cell></row><row><cell>BE</cell><cell>svm</cell><cell>0.599305</cell><cell>+14.9923%</cell></row><row><cell>AT</cell><cell>bert-all-lang</cell><cell>0.598202</cell><cell>+7.8439%</cell></row><row><cell>FR</cell><cell>bert-all-lang</cell><cell>0.57853</cell><cell>+14.9438%</cell></row><row><cell>DK</cell><cell>svm</cell><cell>0.570611</cell><cell>+0.6859%</cell></row><row><cell>IS</cell><cell>bert-all-lang</cell><cell>0.561041</cell><cell>+17.0720%</cell></row><row><cell>HR</cell><cell>bert-all-lang</cell><cell>0.560266</cell><cell>+12.7540%</cell></row><row><cell>IT</cell><cell>bert-all-lang</cell><cell>0.558772</cell><cell>+0.1902%</cell></row><row><cell>NL</cell><cell>bert-all-lang</cell><cell>0.55518</cell><cell>+4.2632%</cell></row><row><cell>FI</cell><cell>svm</cell><cell>0.551095</cell><cell>+0.8286%</cell></row><row><cell>LV</cell><cell>bert-all-lang</cell><cell>0.535344</cell><cell>+8.1636%</cell></row><row><cell>SI</cell><cell>svm</cell><cell>0.532712</cell><cell>+14.1417%</cell></row><row><cell>BA</cell><cell>bert-all-lang</cell><cell>0.526129</cell><cell>+11.0575%</cell></row><row><cell>EE</cell><cell>bert-all-lang</cell><cell>0.525498</cell><cell>+5.0717%</cell></row><row><cell>CZ</cell><cell>baseline</cell><cell>0.518866</cell><cell>+0.0000%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_7"><head>Table 7</head><label>7</label><figDesc>Best Power Evaluation per Parliament</figDesc><table><row><cell cols="4">Parliament Our approach F1 Score F1 Difference to baseline</cell></row><row><cell>HU</cell><cell>svm</cell><cell>0.889968</cell><cell>+3.2326%</cell></row><row><cell>ES-GA</cell><cell>svm</cell><cell>0.880689</cell><cell>+5.1572%</cell></row><row><cell>ES-CT</cell><cell>svm</cell><cell>0.846702</cell><cell>+6.7065%</cell></row><row><cell>TR</cell><cell>logreg</cell><cell>0.836893</cell><cell>+0.5902%</cell></row><row><cell>PL</cell><cell>logreg</cell><cell>0.767386</cell><cell>+0.8908%</cell></row><row><cell>ES-PV</cell><cell>svm</cell><cell>0.742512</cell><cell>+2.7673%</cell></row><row><cell>RS</cell><cell>bertaugm</cell><cell>0.741145</cell><cell>+9.3724%</cell></row><row><cell>GB</cell><cell>logreg</cell><cell>0.722565</cell><cell>+1.1512%</cell></row><row><cell>LV</cell><cell>svm</cell><cell>0.688353</cell><cell>+17.6953%</cell></row><row><cell>BG</cell><cell>baseline</cell><cell>0.681117</cell><cell>+0.0000%</cell></row><row><cell>HR</cell><cell>bert-basic</cell><cell>0.68003</cell><cell>+7.7972%</cell></row><row><cell>AT</cell><cell>logreg</cell><cell>0.666726</cell><cell>+0.6524%</cell></row><row><cell>FR</cell><cell>bert-basic</cell><cell>0.666391</cell><cell>+0.7619%</cell></row><row><cell>GR</cell><cell>bert-basic</cell><cell>0.664713</cell><cell>+3.7555%</cell></row><row><cell>CZ</cell><cell>logreg</cell><cell>0.654461</cell><cell>+1.8956%</cell></row><row><cell>ES</cell><cell>baseline</cell><cell>0.652175</cell><cell>+0.0000%</cell></row><row><cell>IT</cell><cell>bert-basic</cell><cell>0.639566</cell><cell>+21.3734%</cell></row><row><cell>NL</cell><cell>logreg</cell><cell>0.636119</cell><cell>+2.0396%</cell></row><row><cell>DK</cell><cell>bertaugm</cell><cell>0.629958</cell><cell>+6.8779%</cell></row><row><cell>PT</cell><cell>baseline</cell><cell>0.619798</cell><cell>+0.0000%</cell></row><row><cell>BE</cell><cell>baseline</cell><cell>0.612543</cell><cell>+0.0000%</cell></row><row><cell>SI</cell><cell>bertaugm</cell><cell>0.602376</cell><cell>+6.9857%</cell></row><row><cell>BA</cell><cell>svm</cell><cell>0.562684</cell><cell>+10.8886%</cell></row><row><cell>FI</cell><cell>baseline</cell><cell>0.561368</cell><cell>+0.0000%</cell></row><row><cell>UA</cell><cell>bert-basic</cell><cell>0.533312</cell><cell>+7.2237%</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://huggingface.co</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://huggingface.co/saibo/legal-roberta-base</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">https://huggingface.co/meta-llama/Meta-Llama-3-70B</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Overview of Touché 2024: Argumentation Systems</title>
		<author>
			<persName><forename type="first">J</forename><surname>Kiesel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ç</forename><surname>Çöltekin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Heinrich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Alshomary</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">D</forename><surname>Longueville</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Erjavec</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Handke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kopp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ljubešić</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Meden</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Mirzakhmedova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Morkevičius</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Reitis-Munstermann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Scharfbillig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Stefanovitch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wachsmuth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fifteenth International Conference of the CLEF Association (CLEF 2024)</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Mulhem</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Quénot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Schwab</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Soulier</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><forename type="middle">M D</forename><surname>Nunzio</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">G S</forename><surname>De Herrera</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin Heidelberg New York</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Empirical Study of the Model Generalization for Argument Mining in Cross-Domain and Cross-Topic Settings</title>
		<author>
			<persName><forename type="first">A</forename><surname>Alhamzeh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Egyed-Zsigmond</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">E</forename><surname>Mekki</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">E</forename><surname>Khayari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Mitrović</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Brunie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Kosch</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Transactions on Large-Scale Data-and Knowledge-Centered Systems LII</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="103" to="126" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title level="m" type="main">BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding</title>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
		<idno>CoRR abs/1810.04805</idno>
		<ptr target="http://arxiv.org/abs/1810.04805.arXiv:1810.04805" />
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">LEGAL-BERT: The muppets straight out of law school</title>
		<author>
			<persName><forename type="first">I</forename><surname>Chalkidis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fergadiotis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Malakasiotis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Aletras</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Androutsopoulos</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2020.findings-emnlp.261</idno>
	</analytic>
	<monogr>
		<title level="m">Findings of the Association for Computational Linguistics: EMNLP 2020, Association for Computational Linguistics</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="2898" to="2904" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">When Does Pretraining Help? Assessing Self-Supervised Learning for Law and the CaseHOLD Dataset</title>
		<author>
			<persName><forename type="first">L</forename><surname>Zheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Guha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">R</forename><surname>Anderson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Henderson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">E</forename><surname>Ho</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2104.08671</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 18th International Conference on Artificial Intelligence and Law</title>
				<meeting>the 18th International Conference on Artificial Intelligence and Law</meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">Pile of Law: Learning Responsible Data Filtering from the Law and a 256GB Open-Source Legal Dataset</title>
		<author>
			<persName><forename type="first">P</forename><surname>Henderson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">S</forename><surname>Krass</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Guha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Jurafsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">E</forename><surname>Ho</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/2207.00220" />
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">Dropout in Neural Networks</title>
		<author>
			<persName><forename type="first">H</forename><surname>Yadav</surname></persName>
		</author>
		<ptr target="https://towardsdatascience.com/dropout-in-neural-networks-47a162d621d9" />
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">How to Understand Sigmoid Function in Artificial Neural Networks?</title>
		<author>
			<persName><forename type="first">S</forename><surname>Vishwakarma</surname></persName>
		</author>
		<ptr target="https://www.analyticsvidhya.com/blog/2023/01/why-is-sigmoid-function-important-in-artificial-neural-networks/" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<ptr target="https://github.com/meta-llama/llama3/blob/main/MODEL_CARD.md" />
		<title level="m">Llama 3 model card</title>
				<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note>AI@Meta</note>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Trump&apos;s and Clinton&apos;s Style and Rhetoric during the 2016 Presidential Election</title>
		<author>
			<persName><forename type="first">J</forename><surname>Savoy</surname></persName>
		</author>
		<idno type="DOI">10.1080/09296174.2017.1349358</idno>
		<idno>doi:10.1080/09296174.2017.1349358</idno>
		<ptr target="https://doi.org/10.1080/09296174.2017.1349358" />
	</analytic>
	<monogr>
		<title level="j">Journal of Quantitative Linguistics</title>
		<imprint>
			<biblScope unit="volume">25</biblScope>
			<biblScope unit="page" from="168" to="189" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Query Expansion, Argument Mining and Document Scoring for an Efficient Question Answering System</title>
		<author>
			<persName><forename type="first">A</forename><surname>Alhamzeh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bouhaouel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Egyed-Zsigmond</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Mitrović</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Brunie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Kosch</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction</title>
				<editor>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Degli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Esposti</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Sebastiani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Macdonald</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Pasi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Hanbury</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Potthast</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><surname>Ferro</surname></persName>
		</editor>
		<meeting><address><addrLine>Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer International Publishing</publisher>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="162" to="174" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Language Reasoning by means of Argument Mining and Argument Quality</title>
		<author>
			<persName><forename type="first">A</forename><surname>Alhamzeh</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
		<respStmt>
			<orgName>Universität Passau</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Ph.D. thesis</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
