<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Checker Hacker at CheckThat! 2024: Detecting Check-Worthy Claims and Analyzing Subjectivity with Transformers Notebook for the CheckThat! Lab at CLEF 2024</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Syeda</forename><surname>Duae Zehra</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Dhanani School of Science and Engineering</orgName>
								<orgName type="institution">Habib University</orgName>
								<address>
									<settlement>Karachi</settlement>
									<country key="PK">Pakistan</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Kushal</forename><surname>Chandani</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Dhanani School of Science and Engineering</orgName>
								<orgName type="institution">Habib University</orgName>
								<address>
									<settlement>Karachi</settlement>
									<country key="PK">Pakistan</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Muhammad</forename><surname>Khubaib</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Dhanani School of Science and Engineering</orgName>
								<orgName type="institution">Habib University</orgName>
								<address>
									<settlement>Karachi</settlement>
									<country key="PK">Pakistan</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Ahmed</forename><forename type="middle">Ali</forename><surname>Aun Muhammed</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Dhanani School of Science and Engineering</orgName>
								<orgName type="institution">Habib University</orgName>
								<address>
									<settlement>Karachi</settlement>
									<country key="PK">Pakistan</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Faisal</forename><surname>Alvi</surname></persName>
							<email>faisal.alvi@sse.habib.edu.pk</email>
							<affiliation key="aff0">
								<orgName type="department">Dhanani School of Science and Engineering</orgName>
								<orgName type="institution">Habib University</orgName>
								<address>
									<settlement>Karachi</settlement>
									<country key="PK">Pakistan</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Abdul</forename><surname>Samad</surname></persName>
							<email>abdul.samad@sse.habib.edu.pk</email>
							<affiliation key="aff0">
								<orgName type="department">Dhanani School of Science and Engineering</orgName>
								<orgName type="institution">Habib University</orgName>
								<address>
									<settlement>Karachi</settlement>
									<country key="PK">Pakistan</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Checker Hacker at CheckThat! 2024: Detecting Check-Worthy Claims and Analyzing Subjectivity with Transformers Notebook for the CheckThat! Lab at CLEF 2024</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">2068B7AF15EE5AD0314A60BB19EAB006</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T18:00+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>CLEF CheckThat!, fact-checking, transformer models, binary classification, dataset 0009-0002-3207-1826 (S. D. Zehra)</term>
					<term>0009-0002-0954-7742 (K. Chandani)</term>
					<term>0009-0002-0699-4029 (M. Khubaib)</term>
					<term>0009-0004-5910-6018 (A. A. A. Muhammed)</term>
					<term>0000-0003-3827-7710 (F. Alvi)</term>
					<term>0009-0009-5166-6412 (A. Samad)</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper represents our approach on the CheckThat! Lab designed to address the issue of disinformation. We participated in CheckThat! Lab Task 1 which focuses on identifying check-worthy claims in various forms of media, and Task 2 which targets the detection of subjective viewpoints in news articles. For both tasks we focused on the English dataset only. For task 1, after standard preprocessing, we used an ensemble approach where we combined two models, namely BERT-Base-Uncased and XLM-RoBERTa-Base in order to finetune and to find the average probabilities to determine a unified ensemble probability. For task 1 our F1 score was 0.696 and our rank was 14th in the English leaderboard. For task 2 we augmented our data after standard pre-processing using Google AI Studio and it's gemini-1.0-pro-latest model and then used the transformer-based model RoBERTa and finetuned it on the augmented dataset. For task 2, our macro F1 score was 0.7081 and our rank was 4th in the English leaderboard.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The CLEF CheckThat! Lab <ref type="bibr" target="#b17">[17]</ref> initiative is at the forefront of technological developments in automated fact-checking, aiming to combat misinformation in the digital age. Misinformation poses significant risks to public discourse and democratic processes, making the development of effective fact-checking tools crucial. In the 2024 edition <ref type="bibr" target="#b0">[1]</ref> <ref type="bibr" target="#b17">[17]</ref>, the Lab focuses on two key tasks, each addressing critical aspects of this challenge.</p><p>The first one concentrated on assessing the check-worthiness of claims made in tweets and other English texts. This involves identifying which statements require verification, thereby prioritizing efforts, as not all claims can be fact-checked due to resource constraints, and determining check-worthiness ensures that the most impactful misinformation is addressed promptly. The second task aimed to distinguish subjective opinions from objective facts in the sentences of the news articles, something essential for maintaining factual integrity and preventing spread of misinformation. By accurately identifying and separating opinions from facts, we can improve the reliability of news content and support informed public discourse. Unlike sentiment analysis, which would be focused on identifying emotional tones, subjectivity analysis actually aims to improve the working of Task 1 as it aims at discerning statements that may require verification (subjective) from those presenting factual information (objective). By categorizing claims, fact-checkers can prioritize rigorous scrutiny for subjective claims that may influence public opinion or require context evaluation, while focusing factual verification efforts on objective claims backed by evidence. Together, these tasks help in providing a comprehensive approach to validating information, preventing the spread of misinformation, and upholding the credibility of information sources. Both tasks make use of binary classification and measure effectiveness through F1 scores, ensuring precise and efficient validation of information, and preventing the spread of misinformation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Literature Review</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Task 1</head><p>In recent years, the CLEF CheckThat! competition has showcased innovative approaches to claim detection. Top teams have consistently relied on transformer-based models to enhance their systems. Accenture, the top-ranked team in 2020, utilized a RoBERTa-based model, incorporating mean pooling and dropout layers to improve generalization and reduce overfitting <ref type="bibr">[4][5]</ref>. This strategy helped them achieve strong performance over baseline models.</p><p>In 2021, NLP&amp;IR@UNED explored several pre-trained transformer models, discovering that BERTweet was the most effective on the development set. BERTweet, trained on 850 million English tweets and 23 million COVID-19-specific tweets, excelled at identifying check-worthy claims <ref type="bibr">[6][8]</ref>. The second-place team, Fight for 4230, also used BERTweet but added a dropout layer and implemented data augmentation techniques <ref type="bibr" target="#b6">[7]</ref> <ref type="bibr" target="#b7">[8]</ref>. In the following year, PoliMi-FlatEarthers stood out by fine-tuning GPT-3 for Task 1B. They combined deep learning with domain-specific customization to accurately classify check-worthy claims <ref type="bibr" target="#b8">[9]</ref> <ref type="bibr" target="#b9">[10]</ref>. Finally, in 2023, OpenFact leveraged a fine-tuned GPT-3 model, utilizing a rich, annotated dataset of sentences from political debates and speeches. Their data-centric approach, tailored specifically for fact-checking, helped them outperform other submissions <ref type="bibr" target="#b11">[11]</ref> <ref type="bibr" target="#b12">[12]</ref>. Additionally, recent research has shown that models like FACT-GPT, which use synthetic data generated by large language models for training, can closely match human judgment in identifying related claims, highlighting the potential for AI tools to enhance the fact-checking process <ref type="bibr" target="#b20">[20]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Task 2</head><p>This task has only appeared in one previous edition CheckThat! 2023. The top submissions used many different models, the most used were BERT, RoBERTa, ChatGPT and GPT3. Team DWReCo <ref type="bibr" target="#b13">[13]</ref> <ref type="bibr" target="#b16">[16]</ref> got the best score in the English category. There approach involved augmenting the dataset using GPT and then trained on RoBERTa. Two other teams also went with a data augmenting approach. The overall best score on the multilingual dataset was achieved by Team NN <ref type="bibr" target="#b14">[14]</ref>[16] who used the XLMRoBERTa model and trained it on the multilingual dataset. Team Thesis Titan <ref type="bibr" target="#b15">[15]</ref>[16] achieved top positions in 4 languages. Their approach was to train the mDeBERTa model finetuned for each specific languages seperately allowing them to achieve those scores. Many other teams also tried an ensemble approach and got decent results.</p><p>Similar to Team DWReCo's strategy, FACT-GPT utilized large language models (LLMs) to generate synthetic training data, enhancing the adaptability of models for specific tasks, which is crucial for claim matching in fact-checking contexts. Like the approach of using XLMRoBERTa for multilingual datasets, FACT-GPT demonstrated that fine-tuning language models on synthetic datasets could improve classification accuracy and reduce computational costs. Both FACT-GPT and the approaches in CheckThat! 2023 emphasize the importance of leveraging AI to assist and enhance human expertise in the fact-checking process. <ref type="bibr" target="#b20">[20]</ref> 3. Task 1</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Our Approach</head><p>The goal of Task 1 <ref type="bibr" target="#b1">[2]</ref> was to evaluate the necessity of fact-checking claims in tweets and transcriptions. This typically requires either the expertise of professional fact-checkers or answers to several auxiliary questions by human annotators.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.1.">Data Preparation, Model Training and Evaluation</head><p>We were provided with three datasets: <ref type="bibr" target="#b18">[18]</ref> the training dataset, the dev dataset, and the test-dev dataset. Later, we received the fourth dataset, the main test dataset which was unlabeled. Our initial modeling used the following parameters with the BERT-base-uncased model: After training, we used the model to process the test-dev dataset. The procedure involved:</p><p>1. Tokenizing the text entries. 2. Feeding the tokenized data into the model. 3. Converting the output logits to probabilities using a sigmoid function. 4. Classifying each entry as "Yes" or "No" based on a probability threshold of 0.5. 5. Collecting these classifications and their corresponding "Sentence_id" into a list for comparison with the original labels.</p><p>The approach achieved an F1 score of 0.80 on the test-dev dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.2.">Modifications Made For Final Approach</head><p>To improve results, we experimented with various models like Alberta, RoBERTa-base, XLM-RoBERTa, and ELECTRA. The most significant improvement was observed with XLM-RoBERTa-base and BERT-base-uncased. We then implemented an ensemble approach with these two models using the following training configurations:</p><p>• Both trained models were evaluated on the test-dev dataset. Each text data point from the test dataset was processed by both models, and their predictions were averaged to form a single ensemble probability. This probability determined the final label ("Yes" or "No"), which was collected along with the text's unique identifier into a list. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Results</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Our Approach</head><p>The goal of Task 2 was to evaluate the Subjectivity of news articles and decide whether a sentence from the news article <ref type="bibr" target="#b2">[3]</ref>[19] was subjective or objective.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.1.">Data Preparation, Model Training and Evaluation</head><p>Our focus was on the English datasets: the training dataset, the dev dataset, and the test-dev dataset. We used data augmentation to enhance our dataset as the train dataset was very small and the model was not able to learn and effectively. We initially tried to augment the data using WordNet model and the NTLK library. This method changed one word at random from the sentence and replaced it with its synonyms.</p><p>Our initial modeling was done using mDeBERTa and we used the following parameters:</p><p>• After training, we used the model to process the test-dev dataset. The procedure involved:</p><p>1. Processing the data and tokenizing the text entries.</p><p>2. Feeding the tokenized data into the model. 3. Converting the output logits to probabilities. 4. Classifying each entry as "Subj" or "Obj" using Sigmoid and Argmax. 5. Collecting these classifications into a list for comparison with the original labels.</p><p>The approach achieved an F1 score of 0.76. This was achieved on the dataset that had been augmented using the WordNet model and NTLK.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.2.">Modifications Made For Final Approach</head><p>The current approach of changing the words with their synonyms at times did not portray the sentence correctly. We then decided to use the Gemini Api using Google AI Studio and it's 'gemini-1.0-pro-latest' model and augmented our data. The approach used in this case was to create three similar sentences for each of the "Objective" label and five similar sentences for each of the "Subjective" label. This allowed us to have a more balanced dataset and allowed the model to have a better learning. We then imported the dataset called "data", which has been uploaded on GitHub as well. While modifying, we tried different models and even used the ensemble approach using models such as RoBERTa-base, mDeBERTa, RoBERTa-xlm, and BERT-base, but the best results were achieved using RoBERTa-base alone, hence we used that for our final submission using the following training configurations: The probability calculated went through a probability threshold of 0.5, based on which we determined the final label ("Subj" or "Obj"). </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Results</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Analysis</head><p>We saw an overall drop in the scores of our model as it achieved high scores on the training set compared to the dev set, dev-test, and test set scores, which indicates potential overfitting. This means that the model did not perform well with new, unseen data.</p><p>In Task 1, the validation loss showed a slight increase, which potentially contributed to the model's underperformance on new data. This increase in validation loss indicates that the model might have started overfitting to the training data, thereby reducing its generalizability. As a result, when the model was applied to the test set, it did not produce equally good results. Additionally, the class imbalance in the dataset could have affected the model's performance. With fewer instances labeled as check-worthy compared to non-check-worthy, the model might have struggled to accurately identify the check-worthy instances, leading to a lower overall score. The preprocessing steps, while essential for cleaning and preparing the data, might not have fully addressed the inherent variability in the text, further complicating the model's ability to generalize well to unseen data.</p><p>Moreover in task 2, a reason for the low SUBJ F1 score on the test set suggests that the model had difficulty with the "SUBJ" class. One possible reason for this could be that the features used for identifying the "SUBJ" class may not be as strong or distinctive, or there might be more variability or noise in the "SUBJ" class in the test set compared to the training set. Another reason for the low SUBJ F1 could be the way we conducted our data augmentation. The approach that we used, created three similar sentences for each of the "Objective" label and five similar sentences for each of the "Subjective" label. Considering all the sentences might have been similar to the original one from which they were made, we might have experienced over-fitting as the features of those sentences might have been similar.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>In conclusion, our detailed exploration in the CheckThat! Lab 2024 challenge demonstrated the significant capabilities of transformer-based models in tasks of check-worthiness detection and subjectivity analysis. For Task 1, the ensemble method combining XLM-RoBERTa and BERT-base-uncased models effectively navigated the complexities of identifying check-worthy claims. By using a strategic ensemble of predictions and applying a robust training regimen involving multiple epochs (up to 5) and a learning rate of 5 × 10 −5 .</p><p>In Task 2, the fine-tuned RoBERTa model proved better than the other models we had tested as it showed better performance in differentiating the subjective from objective statements on the devtest file, utilizing a refined approach with a lower learning rate (5 × 10 −6 ) and an increased number of epochs <ref type="bibr" target="#b12">(12)</ref>, ensuring thorough learning. However, the performance, as indicated by a macro F1 score of 0.7081 and an F1 score of 0.54 for the SUBJ class, suggests room for improvement. A deeper analysis reveals that the model struggled with the "SUBJ" class, possibly due to weaker feature representation or greater variability and noise in the test set. Another issue that might have been prevalent is of class imabalance which might have led to weaker SUBJ identification. Future work could focus on enhancing the feature set for this class and reducing noise through better data preprocessing and augmentation.</p><p>Data augmentation played a crucial role here, bolstering the dataset and thereby enhancing the model's ability to handle nuanced textual variations. While these results were promising, they also suggest potential areas for further refinement to optimize performance, particularly in handling more complex misinformation scenarios. These efforts exemplify the essential role of adaptive, transformer-based architectures in leveraging deep learning for critical media literacy tasks in a multilingual context.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>•</head><label></label><figDesc>Batch size: 8 for both training and validation • Learning rate: 2 × 10 −5 • Number of epochs: 3</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>Batch size: 16 for both training and validation • Learning rate: 5 × 10 −5 • Number of epochs: 5 • Weight Decay: 0.005</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head></head><label></label><figDesc>Batch size: 16 for both training and validation • Learning rate: 5 × 10 −5 • Number of epochs: 6 • Warmup steps: 100 • Weight decay: 0.01</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>•</head><label></label><figDesc>Batch size: 64 for both training and validation • Learning rate: 5 × 10 −6 • Number of epochs: 12 • Warmup steps: 100 • Weight decay: 0.01</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>Performance metrics for Task 1 across different datasets</figDesc><table><row><cell>Task 1</cell><cell>dev Set</cell><cell>dev-test Set</cell><cell>Test Set</cell></row><row><cell>F1 scores</cell><cell>0.93</cell><cell>0.87</cell><cell>0.696</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 :</head><label>2</label><figDesc>Performance Metrics for Task 2 across different datasets</figDesc><table><row><cell>Task 2</cell><cell>Dev Set</cell><cell>Dev-test Set</cell><cell>Test Set</cell></row><row><cell>MACRO F1</cell><cell>0.86</cell><cell>0.82</cell><cell>0.708</cell></row><row><cell>SUBJ F1</cell><cell>0.82</cell><cell>0.83</cell><cell>0.54</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>The authors would like to acknowledge the support provided by the Office Of Research (OoR) at Habib University, Karachi, Pakistan for funding this project through internal research grant IRG-2235.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">The CLEF-2024 CheckThat! Lab: Check-Worthiness, Subjectivity, Persuasion, Roles, Authorities, and Adversarial Robustness</title>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-031-56069-9_62</idno>
		<ptr target="https://doi.org/10.1007/978-3-031-56069-9_62" />
	</analytic>
	<monogr>
		<title level="m">Advances in Information Retrieval. ECIR 2024</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">N</forename><surname>Goharian</surname></persName>
		</editor>
		<meeting><address><addrLine>Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2024">2024</date>
			<biblScope unit="volume">14612</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Overview of the CLEF-2024 CheckThat! Lab Task 1 on Check-Worthiness Estimation of Multigenre Content</title>
		<author>
			<persName><forename type="first">M</forename><surname>Hasanain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Suwaileh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Weering</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Caselli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zaghouani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Alam</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2024 -Conference and Labs of the Evaluation Forum, CLEF 2024</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>García Seco De Herrera</surname></persName>
		</editor>
		<meeting><address><addrLine>Grenoble, France</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Overview of the CLEF-2024 CheckThat! Lab Task 2 on Subjectivity in News Articles</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Struß</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Ruggeri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Alam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dimitrov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Galassi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Siegel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wiegand</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2024 -Conference and Labs of the Evaluation Forum, CLEF 2024</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>García Seco De Herrera</surname></persName>
		</editor>
		<meeting><address><addrLine>Grenoble, France</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Accenture at CheckThat! 2020: If you say so: Post-hoc fact-checking of claims using transformer-based models</title>
		<author>
			<persName><forename type="first">E</forename><surname>Williams</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rodrigues</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Novak</surname></persName>
		</author>
		<editor>Cappellato et al.</editor>
		<imprint>
			<date type="published" when="2020">2020</date>
			<pubPlace>CLEF</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Overview of CheckThat! 2020 English: Automatic Identification and Verification of Claims in Social Media</title>
		<author>
			<persName><forename type="first">S</forename><surname>Shaar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Nikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Babulkov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Alam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Elsayed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hasanain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Suwaileh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Haouari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
		<editor>Cappellato et al.</editor>
		<imprint>
			<date type="published" when="2020">2020</date>
			<pubPlace>CLEF</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>-R. Juan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Martinez-Rico</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Araujo</surname></persName>
		</author>
		<title level="m">NLP&amp;IR@UNED at CheckThat! 2021: Checkworthiness estimation and fake news detection using transformer models</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">X</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Fung</surname></persName>
		</author>
		<title level="m">Fight for 4230 at CLEF CheckThat! 2021: Domain-specific preprocessing and pretrained model for ranking claims by check-worthiness</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Overview of the CLEF-2021 CheckThat! Lab: Task 1 on Check-Worthiness Estimation in Tweets and Political Debates</title>
		<author>
			<persName><forename type="first">S</forename><surname>Shaar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hasanain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Hamdan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><forename type="middle">S</forename><surname>Ali</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Haouari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Nikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kutlu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">S</forename><surname>Kartal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Alam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Miguez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Beltrán</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Elsayed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">CLEF</title>
		<editor>Cappellato et al.</editor>
		<imprint>
			<biblScope unit="page" from="369" to="392" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">PoliMi-FlatEarthers at CheckThat! 2022: GPT-3 applied to claim detection</title>
		<author>
			<persName><forename type="first">S</forename><surname>Agrestia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Hashemianb</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Carmanc</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2022 -Conference and Labs of the Evaluation Forum, CLEF &apos;2022</title>
				<meeting><address><addrLine>Bologna, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Overview of the CLEF</title>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Alam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Míguez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Caselli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kutlu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zaghouani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Shaar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Mubarak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Nikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">S</forename><surname>Kartal</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">CheckThat! Lab: Task 1 on Identifying Relevant Claims in Tweets</title>
	</analytic>
	<monogr>
		<title level="j">CLEF</title>
		<editor>Cappellato et al.</editor>
		<imprint>
			<biblScope unit="page" from="368" to="392" />
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">-head gpt vs. bert -a comparative study of transformers language models for the detection of check-worthy claims</title>
		<author>
			<persName><forename type="first">M</forename><surname>Sawiński</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Wecel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Ksieżniak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Stróżyna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Lewoniewski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Stolarski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Abramowicz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Openfact at CheckThat! 2023: Head-to</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Overview of the CLEF-2023 CheckThat! Lab: Task 1 on Check-Worthiness in Multimodal and Multigenre Content</title>
		<author>
			<persName><forename type="first">F</forename><surname>Alam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">S</forename><surname>Cheema</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">K</forename><surname>Shahi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Hakimov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hasanain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Míguez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Mubarak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zaghouani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">CLEF</title>
		<editor>Cappellato et al.</editor>
		<imprint>
			<biblScope unit="page" from="219" to="235" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">B</forename><surname>Schlicht</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Khellaf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Altiok</surname></persName>
		</author>
		<title level="m">Dwreco at CheckThat! 2023: Enhancing subjectivity detection through style-based data sampling</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<author>
			<persName><forename type="first">K</forename><surname>Dey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Tarannum</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Hasan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">R H</forename><surname>Noori</surname></persName>
		</author>
		<title level="m">Nn at CheckThat! 2023: Subjectivity in news articles classification with transformer based models</title>
				<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title level="m" type="main">Thesis titan at CheckThat! 2023: Language-specific fine-tuning of mdebertav3 for subjectivity detection</title>
		<author>
			<persName><forename type="first">F</forename><surname>Leistra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Caselli</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Overview of the CLEF-2023 CheckThat! Lab: Task 2 on Subjectivity Detection</title>
		<author>
			<persName><forename type="first">A</forename><surname>Galassi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Ruggeri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Alam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Caselli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kutlu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Struß</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Antici</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hasanain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Köhler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Korre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Leistra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Muti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Siegel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">D</forename><surname>Türkmen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wiegand</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zaghouani</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">CLEF</title>
		<editor>Cappellato et al.</editor>
		<imprint>
			<biblScope unit="page" from="236" to="249" />
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Overview of the CLEF-2024 CheckThat! Lab: Check-Worthiness, Subjectivity, Persuasion, Roles, Authorities, and Adversarial Robustness</title>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Alam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Struß</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Chakraborty</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Elsayed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Przybyła</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Caselli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Haouari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Piskorski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Ruggeri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Suwaileh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fifteenth International Conference of the CLEF Association</title>
				<editor>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Mulhem</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Quénot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Schwab</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Soulier</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><forename type="middle">M</forename><surname>Di Nunzio</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>García Seco De Herrera</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<meeting><address><addrLine>CLEF</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2024">2024. 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Fighting the COVID-19 Infodemic: Modeling the Perspective of Journalists, Fact-Checkers, Social Media Platforms, Policy Makers, and the Society</title>
		<author>
			<persName><forename type="first">F</forename><surname>Alam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Shaar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Dalvi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Sajjad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Nikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Mubarak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Abdelali</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Durrani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Darwish</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Findings of the Association for Computational Linguistics: EMNLP 2021</title>
				<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="611" to="649" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">On the Definition of Prescriptive Annotation Guidelines for Language-Agnostic Subjectivity Detection</title>
		<author>
			<persName><forename type="first">F</forename><surname>Ruggeri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Antici</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Galassi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Korre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Muti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<ptr target="CEUR-WS.org" />
	</analytic>
	<monogr>
		<title level="m">Text2Story@ECIR, CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="volume">3370</biblScope>
			<biblScope unit="page" from="103" to="111" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">FACT-GPT: Fact-Checking Augmentation via Claim Matching with LLMs</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">C</forename><surname>Choi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Ferrara</surname></persName>
		</author>
		<idno type="DOI">10.1145/3589335.3651504</idno>
		<ptr target="https://doi.org/10.1145/3589335.3651504" />
	</analytic>
	<monogr>
		<title level="m">Companion Proceedings of the ACM on Web Conference</title>
				<imprint>
			<date type="published" when="2024-05-13">2024. 13 May 2024</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
