<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">AuthEv-LKolb at CheckThat! 2024: A Two-Stage Approach To Evidence-Based Social Media Claim Verification</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Luis</forename><surname>Kolb</surname></persName>
							<email>kolb.luis@gmail.com</email>
							<affiliation key="aff0">
								<orgName type="department">Data Science Research Unit</orgName>
								<orgName type="institution">TU Wien</orgName>
								<address>
									<addrLine>Favoritenstraße 9-11/194-04</addrLine>
									<postCode>1040</postCode>
									<settlement>Vienna</settlement>
									<country key="AT">Austria</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Allan</forename><surname>Hanbury</surname></persName>
							<email>allan.hanbury@tuwien.ac.at</email>
							<affiliation key="aff0">
								<orgName type="department">Data Science Research Unit</orgName>
								<orgName type="institution">TU Wien</orgName>
								<address>
									<addrLine>Favoritenstraße 9-11/194-04</addrLine>
									<postCode>1040</postCode>
									<settlement>Vienna</settlement>
									<country key="AT">Austria</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">AuthEv-LKolb at CheckThat! 2024: A Two-Stage Approach To Evidence-Based Social Media Claim Verification</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">58049E83C7BC89F212802A731C7C945D</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:55+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>fact-checking</term>
					<term>natural language processing</term>
					<term>information retrieval</term>
					<term>CLEF 2024</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper covers our submission to CLEF 2024 CheckThat! Lab task 5: Authority Evidence for Rumor Verification. Misinformation as claims on social media platforms is an ever present issue. We present a two-stage approach to verify claims posted on social media based on evidence posted by authority accounts to the same platform. We conduct experiments to find the optimal setup with respect to the target metrics specified in the CLEF 2024 CheckThat! Lab, where we are participating in Task 5. Our experiments show that Large Language Models, of which we compare GPT-4 and Llama3-70B, are suited to this particular verification task. The paper finally presents areas where further improvements can be explored.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>This paper covers our submission to CLEF 2024 CheckThat! Lab task 5: Authority Evidence for Rumor Verification. The descriptions for all tasks, including our own, are provided in the conference paper by the lab organizers <ref type="bibr" target="#b0">[1]</ref>.</p><p>There are many options available to large platform operators to combat misinformation on their platforms, like professional fact-checking services or even manually fact-checking claims on their platform. Manually checking every reported post has turned into a task that is no longer a viable option for most large platforms, due to the sheer volume of content uploaded by users. There are improvements to these methods that platforms can implement, like identifying and matching similar claims to already fact-checked claims and reusing the work that already went into fact-checking the original claim. This was already a task at the CLEF 2022 CheckThat! Lab <ref type="bibr" target="#b1">[2]</ref>. However, there are also alternative approaches.</p><p>Specifically on X.com (formerly Twitter), a community fact-checking system is in place, colloquially called "Community Notes". In a 2022 study, Pröllochs <ref type="bibr" target="#b2">[3]</ref> investigated the impact of this feature, and one of the findings was that the feature's "[...] community-driven approach faces challenges concerning opinion speculation and polarization among the user base -in particular with regards to influential user accounts. " (p.11 <ref type="bibr" target="#b2">[3]</ref>)</p><p>In this paper, we present a more automated approach to fact-checking claims on social media, using official government statements on the same platform to verify claims, which could be used both as a stand-alone service, and as a tool to assist human fact-checkers and fact-checking services. There are some drawbacks to relying on official authority accounts rather than the community, which are discussed in Section 5. <ref type="bibr" target="#b2">3</ref>.</p><p>The official CLEF 2024 CheckThat! Lab Task 5 is defined as: "Given a rumor expressed in a tweet and a set of authorities (one or more authority Twitter accounts) for that rumor, represented by a list of tweets from their timelines during the period surrounding the rumor, the system should retrieve up to 5 evidence tweets from those timelines, and determine if the rumor is supported (true), refuted (false), or unverifiable (in case not enough evidence to verify it exists in the given tweets) according to the evidence" <ref type="bibr" target="#b3">[4]</ref>.</p><p>We experimented with several setups and combinations of different strategies. Our approach involved and tested two stages: a retrieval stage, and a verification stage.</p><p>• In the retrieval stage, for a given claim (also referred to as "rumor"), we aim to retrieve evidence from the set of all tweets relevant to that claim. • In the verification stage, we use the retrieved evidence to predict a label for the claim (REFUTES, SUPPORTS or NOT ENOUGH INFO).</p><p>We structure our paper into the following sections: Section 2 introduces the data we are working with, and the target measures we use to evaluate our experiment results. Chapter 3 discusses the main objectives of the experiments we conducted during our participation, while Chapter 4 presents our approach to the task. The results of our experiments are presented in Chapter 5. Finally, Chapter 6 concludes our paper, and presents questions and topics for further research.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Task Dataset and Evaluation Measures</head><p>The data we are working with consists of various tweet texts. For every tweet making a claim, there is a set of tweets authored by authority sources, only some of which are relevant to the claim. Notably, the tweet texts do include links to attached images that were posted with the tweet, which could contain some additional information (see Section 6 discussing future work). The only metadata directly included in the dataset is the username and the tweet ID (which can be used to fetch more metadata from the twitter API), but these are present only for authority statements, not for claims (which are only a single "text string").</p><p>Here is an example of what a rumor to be verified looks like: every rumor is provided as JSON, with an ID, the claim text, and a list of statements, each of which contains the account URL that tweeted the statement, the tweet ID, and the tweet text. Labeled data also includes which of the statements are relevant to the claim. The dataset consists of 160 rumors overall, 128 of which were available with ground truth. Our approach did not involve learning-to-rank <ref type="bibr" target="#b4">[5]</ref>. So, in our case, the dataset size is only relevant insofar as a larger and more diverse dataset is likely to cover a wider range of scenarios and topics, on which the proposed system could be tested.</p><p>For this task, the following target measures are considered when evaluating performance, as specified in the Task description <ref type="bibr" target="#b5">[6]</ref>:</p><p>• Macro-F1 as the primary measure for the overall verification performance, which averages the F1 score for each of the three labels to account for class imbalance. • Strict Macro-F1 as a secondary measure for verification performance, which additionally considers found evidence. For this measure, a "true positive" needs to have a correct label for the rumor, as well as an overlap of at least one piece of evidence between ground-truth evidence and found evidence. More overlap does not increase the Strict F1 score. • MAP (Mean Average Precision) as the primary measure for retrieval effectiveness.</p><p>• R@5 (Recall at 5 "items", as the system should retrieve at most 5 statements) as a secondary measure for retrieval effectiveness.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Experiment Design</head><p>For this paper, we ran a set of experiments, the results of which are presented in Section 5. We also made a submission to the CheckThat! Lab. This submission and the experiments are separate. The experiments serve to evaluate the effectiveness of different configurations for our proposed setup, and the submission is created using three of those configurations, since each team could submit up to three runs to the CheckThat! Lab. Our proposed setup is illustrated in Figure <ref type="figure" target="#fig_0">1</ref>. This setup was selected based on initial experimentation with a setup oriented after a paper on "Stance Detection" by Haouari et al. <ref type="bibr" target="#b6">[7]</ref>, which was refined over the course of development. We narrowed down the methods we used in initial experiments to the methods described in Section 4.</p><p>In the following list, each number refers to a numbered component in Figure <ref type="figure" target="#fig_0">1</ref>. Our experiments quantitatively evaluate:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Preprocessing and Data Augmentation</head><p>Preprocessing approaches and strategies to combine the individual predictions were also part of our experiments. We aim to obtain the best performing setup of preprocessing strategies for both the retrieval and verification steps, and the best scoring strategies for the verification step.</p><p>Prepocessing and data augmentation (Figure <ref type="figure" target="#fig_0">1</ref> component number 1) are optional features:</p><p>• Data augmentation adds the Twitter display name and/or the Twitter author bio to the statement text, depending on the configuration. • Preprocessing cleans up the text: remove line breaks, some special characters like quotes and hashtags, URLs, the pattern "RT @&lt;username&gt;" (added by the Twitter API for quote tweets) and emojis.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Retrieval Stage</head><p>For the retrieval stage (Figure <ref type="figure" target="#fig_0">1</ref> component number 2), there are multiple viable options. The main retrieval methods we focused on (and ran experiments for) were:</p><p>• PyTerrier to create a simple BatchRetrieve pipeline of (BM25) » (PL2). See the PyTerrier documentation for details.<ref type="foot" target="#foot_0">1</ref> • Cosine distances between embeddings obtained through the OpenAI Embeddings API.</p><p>These methods are rather simple, but effective enough for this task. So long as a single relevant source is retrieved, the powerful LLM in the verification stage is able to correctly predict the judgment.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Verification Stage</head><p>For the verification stage (Figure <ref type="figure" target="#fig_0">1</ref> component number 3), we experimented with two major transformerbased approaches:</p><p>• A fine-tuned version of BART (specifically bart-large-mnli, available on Hugging Face), a sequenceto-sequence autoencoder by Facebook (Meta AI) <ref type="bibr" target="#b7">[8]</ref>, for zero-shot Natural Language Inference (NLI) <ref type="bibr" target="#b8">[9]</ref>. <ref type="foot" target="#foot_1">2</ref> We initially used NLI models for the verification stage, which use a classification approach to classify a combined text input as ENTAILMENT or CONTRADICTION. However the deep natural language understanding of LLMs allows the model to navigate somewhat complex reasoning tasks very effectively. In fact, they outperform the NLI models we used by a wide margin. • Large Language Models (LLMs), for the submission GPT-4 by OpenAI <ref type="bibr" target="#b9">[10]</ref> (specifically the model version named GPT-4-1106-preview), since it performed the best of all available models at the time of the submission period. Due to the relatively low complexity of the reasoning (which is somewhat similar to natural language inference) in the verification step, we theorize that most sufficiently large LLMs would perform similarly (e.g. Llama3-400b or Claude Opus). At the scale of Llama3-70b we saw significant performance drops compared to GPT-4-Turbo.</p><p>Our CheckThat! Lab submission was created using gpt-4-1106-preview as the LLM. After uploading the submission, we re-ran our experiment setup, for which we used gpt-4o-2024-05-13, as it was cheaper and faster to run, and was the newest available LLM from OpenAI.</p><p>For OpenAI completions, we invoke the LLM by using the OpenAI Assistants API, with each claimevidence pairing creating a new thread, and the system prompt being set in the assistant. <ref type="foot" target="#foot_2">3</ref> For the assistant configuration, we used a temperature of 0.01 and a top-p of 0.5. These values should encourage consistent responses. <ref type="foot" target="#foot_3">4</ref> Llama3 completions are obtained using the Hugging Face Inference API with the default parameter values and the model Llama3-70B-Instruct, as more powerful Llama3 models are not yet available. <ref type="foot" target="#foot_4">5</ref>The LLM is prompted in the verification stage with a prompt template that is populated with both the claim and the authority statement, and instructed using the system prompt to adhere to the output format, and to only use information from the prompt, and not to use its domain knowledge or knowledge from training data. The LLM will predict not only a label, but also a confidence in the label between 0 and 1, which will be used to combine the pairwise labels in the next step.</p><p>The system prompt, which gives the LLM instructions it must adhere to, is shown below. The OpenAI LLM Assistant API always adhered to this system prompt during our experimentation. We also activated the "JSON-mode" in the Assistants configuration, which ensures answers follow the format specified in the system prompt, though the system prompt on its own would likely be effective enough to ensure this behavior.</p><p>You are a helpful assistant doing simple reasoning tasks. You will be given a statement and a claim. You need to decide if a statement either supports the claim ("SUPPORTS"), refutes the claim ("REFUTES"), or if the statement is not related to the claim ("NOT ENOUGH INFO"). USE ONLY THE STATEMENT AND THE CLAIM PROVIDED BY THE USER TO MAKE YOUR DECISION. You must also provide a confidence score between 0 and 1, indicating how confident you are in your decision. You must format your answer in JSON format, like this:</p><formula xml:id="formula_0">{"decision": ["SUPPORTS"|"REFUTES"|"NOT ENOUGH INFO"], "confidence": [0...1]} No yapping.</formula><p>Below is a real input message to the LLM (primed with the previous system prompt). In this example, the data was preprocessed and had no external data added:</p><p>"Statement from Authority Account 'LBpresidency': ''The Information Office of the Presidency of the Republic denies a false news broadcast by the MTV station about Baabda Palace preparing a decree naturalizing 4 000 people and recalls that it had denied yesterday the false information published by the French magazine 'Liberation' about the same fabricated news ''" Claim: "Naturalization decree in preparation: Lebanese passports for sale !" Since we score every combination of rumor and evidence separately, we have to combine them to produce an overall label prediction. As part of our experiments, we tested (Figure <ref type="figure" target="#fig_0">1</ref> component number 4):</p><p>• Weighting ("scaling") prediction confidence scores by retrieval score. The retrieval stage, in addition to the top-5 documents, also returns the associated score used to compute the ranking, which optionally can be used here. • Normalizing retrieval scores, as different retrieval systems return retrieval scores on different scales. • Including versus ignoring NOT ENOUGH INFO predictions for the final label score calculation.</p><p>Once we have obtained label predictions and for every claim-statement pairing, we weigh the confidence the LLM in the verification stage predicted using the retrieval score (if this feature is set active in the configuration), and then calculate the mean of the predicted scores (confidences) to obtain our overall label prediction. If the summed, averaged scores cross a significance threshold, we predict the respective SUPPORTS or REFUTES label. The threshold is not tuned or learned, rather it is set manually at 0.15 such that two opposing predictions of roughly equal confidence cancel each other out, unless one predictions is much more significant than the other, opposing prediction. Thus, the threshold accounts for some variation between two roughly equally strong predictions. Our experiments show that for this data set, this simple approach of combining predictions is sufficient.</p><p>Since SUPPORTS predictions are positive, and REFUTES predictions are negative, taking the mean of the two predictions scores emulates a voting system with votes being weighted by the prediction confidences. In this system, NOT ENOUGH INFO predictions do not contribute to the final overall label, as a NOT ENOUGH INFO prediction from the LLM does not indicate any leaning toward either SUPPORTS or REFUTES. Optionally, we include the NOT ENOUGH INFO prediction in the average, lowering the total overall score -potentially below the significance threshold.</p><p>This type of task is related to "stance detection" of authorities, which was introduced in a paper by Haouari et al. (who are also the organizers of the 2024 CheckThat! Lab task 5) in 2023 <ref type="bibr" target="#b6">[7]</ref>. Our approach follows up on their paper, and expands the implementation to also retrieve evidence from a predefined dataset. Graves <ref type="bibr" target="#b10">[11]</ref> lists three families of approaches to automatic fact verification, one of which is "[...] consulting authoritative sources" <ref type="bibr" target="#b10">[11]</ref>. Manually consulting a third-party authority is definitely a valid tool for in-depth fact-checkers, and our system aims to assist these fact-checkers by finding statements an authoritative source already posted publicly, and predicting the stance of the source to the rumor or claim.</p><p>The resources used during development and for the submission are listed here:</p><p>• For OpenAI embeddings and GPT-4-Turbo completions we used the OpenAI API. <ref type="foot" target="#foot_5">6</ref>• Llama3-70b completions were obtained from the Hugging Face "Inference for Pros" API. <ref type="foot" target="#foot_6">7</ref>• BM25, PySerini and TF-IDF retrieval methods as well as bart-large-mnli for verification were computationally cheap enough to effectively run on a local desktop PC (AMD Ryzen 5, Nvidia GTX 970, 16GB memory).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Results and Discussion</head><p>We participated in the CheckThat! Lab Task 5 <ref type="bibr" target="#b3">[4]</ref>, and independently ran experiments to find the best configuration options for our approach. The results of each are reported here, in their own subsections.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Experiment Results</head><p>To test the various configuration options we created, we ran automated experiments on the dev split of the dataset (containing 32 rumors to be verified using the included timeline) to answer this set of research questions:</p><p>• To what extent can tweets ("evidence") relevant to a claim be retrieved from timelines of authority accounts, given an initial claim, a set of authority accounts and the timelines of those authority accounts? • To what extent can a claim, given a list of tweets ("evidence"), accurately be identified as being supported by the evidence (true), refuted by the evidence (false), or unverifiable (not enough evidence to verify it)? • To what extent can a pipeline combining the approaches from RQ1 and RQ2 refute or support a claim, automatically retrieving evidence from the timelines of authority accounts? For the experiments we performed, which are presented in Table <ref type="table">1</ref> to find the best retrieval configuration, given the features we tested (preprocessing, adding author name and author bio), we did not find significant differences looking only at the retrieval evaluation. Generally, the best MAP performance was obtained by the simple PyTerrier retrieval method of scoring with BM25, then re-ranking using PL2 (divergence-from-randomness), using preprocessing and including the author bio in the statement text. It seems that preprocessing slightly improves retrieval performance overall. For the secondary measure, Recall@5, PyTerrier also performed the best.</p><p>In our approach, we ran the experiments to optimize the system for the use case of verification, as a "pipeline" from start to finish (claim and timeline input, to overall label with evidence output). Table <ref type="table" target="#tab_1">2</ref> lists changes in score when a feature is actively used in a configuration, versus when it is not. It also shows the score difference between experiments that used LLAMA3 and those that used GPT-4, and changes in verification score between experiments with each retrieval method described above. The features that were tested are described in Section 4. In Table <ref type="table" target="#tab_1">2</ref>, "Ignore NEI" means ignore NOT ENOUGH INFO (NEI) predictions for the overall score.</p><p>Since we ran experiments in all possible permutations of our configuration options, we calculate the mean score of every configuration where a feature is used, and do the same for every configuration where it is not used (for example, Preprocessing "True" vs. "False", or in the case of Verification methods "OPENAI" vs. "PyTerrier"). The difference in average score gives an indication of the score impact of the feature value. A positive score difference in Table <ref type="table" target="#tab_1">2</ref> means the average score of the configurations using value option 1 was higher than those using value option 2. In most cases, the difference is not meaningful.</p><p>Using GPT-4 over LLAMA3 yields the highest performance gain on average, a noticeable Macro-F1 score increase of about 0.2. This is not surprising, as the GPT-4 model is much more powerful, as mentioned previously. It would be interesting to see score differences on other comparably large language models, like Claude Opus or Google Gemini. However, that comparison is outside the scope of this paper.</p><p>Additionally, it would be interesting to see the influence of different retrieval methods on the verification performance. In our experiments, the difference between retrieval methods is rather small. As mentioned previously, the LLMs in the verification stage are powerful enough that a single piece of relevant evidence usually suffices to predict the correct label. Running the experiments on a more diverse dataset with different retrieval methods in a larger search space might hinder the verification stage from functioning properly. If no relevant evidence is found, the system is likely to predict NOT ENOUGH INFO -as it should.</p><p>The best performing configurations (at rank 1 and 2, see Table <ref type="table">5</ref> in the Appendix) of the system yielded the best results when not using any preprocessing. Preprocessing nearly always removes some amount of signal along with the noise in the data, which might hurt LLM performance more than in helps. Roughly two thirds of all configurations achieving the highest scores used no preprocessing. Overall, the mean Macro-F1 score of system configurations using preprocessing is lower by 0.0139 in our experiments, see Table <ref type="table" target="#tab_1">2</ref>.</p><p>Proposed features like scaling by retrieval score, normalizing retrieval score to [0...1] and including external data did not have a significant impact in our experiments with this dataset. The impact of excluding NOT ENOUGH INFO predictions is noticeable, since in our configuration, the final label is created by averaging the confidences of all pairwise predictions by the LLM, and if the average over that list passes a threshold, a REFUTES or SUPPORTS label is predicted. Including NOT ENOUGH INFO predictions with a value of 0 simply lowers the average score, which at a retrieval-k of 5 pairs can be significant enough to make a difference. In this case, including the NOT ENOUGH INFO predictions in the average score results in the system being presumably too cautious to perform adequately.</p><p>In some cases, the verifier fails to correctly classify SUPPORTS or REFUTES. During our testing, in each of those cases, the system predicts NOT ENOUGH INFO overall, which is the ideal fail case. The system never predicted overall SUPPORTS where the actual overall label is REFUTES, or the other way around.</p><p>See the Appendix for the full tables, or view the Jupyter Notebook with the full tables on GitHub.<ref type="foot" target="#foot_7">8</ref> </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">CLEF Submission Results</head><p>In the CheckThat! Lab Task 5, we participated in the challenge for the English dataset. The measures reported by the Lab organizers were MAP and R@5 for retrieval, and Macro-F1 and Strict Macro-F1 scores for verification (see also Section 2). There was a limit of 3 runs able to be submitted per team, only one of which was allowed to use external data not included in the dataset (our run labeled "secondary1" used author display name and author bio from Twitter, if available). The CheckThat! Lab organizers also provide a baseline score. Here, we report our own results, and this baseline, the full leaderboard is available on the CheckThat! Lab Task 5 website. <ref type="foot" target="#foot_8">9</ref>We submitted three runs, each with different configurations for our setup:</p><p>• "primary": No external data and not preprocessing, only OpenAI embeddings with "raw" data.</p><p>• "secondary1": OpenAI embeddings for retrieval, with external Twitter data about the author added, and no preprocessing. • "secondary2": PyTerrier retrieval method, using preprocessed data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 3</head><p>Selected results for the English retrieval leaderboard. For each other team, the best submission score is presented here.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Team</head><p>Run Label MAP R@5 Retrieval Preprocess External Data IAI Group secondary1 0.628 0.676 bigIR primary 0.604 0.677 Axolotl primary 0.566 0.617 Team DEFAULT primary 0.559 0.634 AuthEv-LKolb (ours) primary 0.549 0.587 Embeddings False False AuthEv-LKolb (ours) secondary2 0.524 0.563 PyTerrier True False AuthEv-LKolb (ours) secondary1 0.510 0.619 Embeddings False True (baseline) 0.335 0.445</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 4</head><p>Selected results for the English verification leaderboard. For all runs, we used GPT-4 as the verification component.</p><p>For each other team, the best submission score is presented here. All three runs used GPT-4 as the verification stage as described in Section 4. Preprocessing and external data are described in Section 4. The configuration options for the combination of the pairwise predictions were all set to "False", meaning no scaling or weighting using the retrieval score, and NOT ENOUGH INFO predictions being included in the average used to calculate the overall label.</p><p>In the retrieval stage, presented in Table <ref type="table">3</ref>, the best score our system achieved was a MAP of 0.549 using the primary run setup. Notably, we achieved a R@5 score of 0.619 using the secondary setup with external data, which would have been 4th on the leaderboard if R@5 was the targeted measure. The highest score was achieved by team "IAI Group", with a MAP of 0.628, who used a "Crossencoder" approach, according to their Run ID on the official leaderboard.</p><p>In the verification stage, we achieved the best result with a Macro-F1 of 0.895 in the secondary system using external data (authority display name and authority bio, obtained from Twitter). Our results can be seen in Table <ref type="table">4</ref>.</p><p>As the leaderboard results show, our approach to retrieval did not work particularly well in comparison to the other participants. However, our verification component significantly outperformed the other participants. Presumably, this demonstrates the strength of Large Language Models in this type of task, where few relevant pieces of evidence are needed to predict correctly, and irrelevant evidence does not introduce significant noise to the overall prediction. Thus, even though our retrieval component was comparatively weaker, the relatively high Recall resulted in good predictions overall.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.">Limitations of the Approach</head><p>There are a few caveats to this proposed setup, and its utility would likely lie in serving as an additional tool in the toolbox used to combat the spread of misinformation. These caveats are:</p><p>• Human fact-checking by neutral sources will most likely be more precise, more reliable and more trusted, assuming the fact-checkers themselves are seen as neutral and trustworthy (which is influenced by a multitude of factors, as analyzed in the study by <ref type="bibr">Primig [12]</ref>).</p><p>• In contrast to "traditional" fact-checking, our approach does not verify the actual truth content in a claim, only whether authority sources support or dispute a claim. For this paper, we are working with the definition of "authority" laid out by Haouari et al. in their 2023 paper <ref type="bibr" target="#b12">[13]</ref>. Authority sources can be government accounts, for example, a Ministry of Health in a given state could be considered an authority related to rumors or claims about public health related matters in the same state. In general, authorities are considered experts in a given area, but not all experts are necessarily considered authorities. Additionally, an account is considered an authority when a rumor is about the account holder themselves. For example, the dev split of the data set contains a rumor about a journalist being involved in a deadly car crash, and the statement "My loved ones and my people who were busy with me: I am fine [...]" posted by the journalist's account is considered authority evidence refuting the rumor. Because of examples like this, we included experiments with adding external data like account name and account bio to the data set. • Another consideration is model selection. For our submission, we used GPT-4-1106-preview, the most recent OpenAI model available at the time. It is important to note that closed models are subject to frequent changes, and "open" models like Llama3-400b should produce more predictable output over longer periods of time. <ref type="foot" target="#foot_9">10</ref> Additionally, closed models are usually subject to content moderation, which could plausibly impact system performance and reliability -the area of fact-checking often deals with controversial claims and statements, after all. Unfortunately, Llama3-400b was not yet publicly available at the time of writing.</p><p>Running the system in the best-performing configuration is also the most expensive way to run the system (both in terms of computing time and API costs). There are possible trade-offs, as if deployment "at scale" or "in production" is desired, some compromises could be necessary:</p><p>• BM25 performs similarly to the OpenAI embeddings cosine distances method. It is also much cheaper to execute, as it does not require an external API call and the associated tokens. • For the best performing configuration, external data is included in, or rather, added to the data set. This inclusion slightly improves performance, but also requires another API call, which is also expensive, due to the restructuring of the X.com (formerly Twitter) API. <ref type="foot" target="#foot_10">11</ref>A recent study by Primig <ref type="bibr" target="#b11">[12]</ref> from 2022 looked at the perception of fact-checkers and fact-checking services in the study population. The author found that, while higher trust in media correlates with trust in fact-checking, there is a significant part of the population who view fact-checking services as propaganda tools of the established government. To increase trust in the system, its purpose needs to be clearly stated: which is to assist users in verifying rumors using official sources. Those users who distrust and reject official sources out of hand will not find the information provided by our system to be helpful.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion and Perspectives for Future Work</head><p>In this paper, we have demonstrated the ability of our proposed setup to generally accurately classify whether official sources SUPPORT or REFUTE unseen rumors in a zero-shot fashion, using the data provided by the task organizers. In a real-world application, some considerations would have to be made with respect to operational aspects like computation costs, as LLMs are expensive to use "at scale". Model selection could also have a significant impact (especially "closed-source" models), as discussed in Section 5.3 .</p><p>In future work, improvements can be made, and extensions of the system need to be checked for performance improvements. Intuitive areas for further experimentation and development are:</p><p>• Do different embedding models for retrieval influence the performance of the verification stage? Do they significantly influence distribution of answers (for example, are there less or more NOT ENOUGH INFO predictions using another embedding model)? • How can the retrieval stage be improved? Retrieval is essential for any fact-checking system to be able to judge a claim, as the verification stage relies on relevant evidence. • How well does the system generalize to other domains and social media platforms? The datasets used were mainly focused on a specific geographical region, auto-translated from Arabic, and the topics of the statement-claim pairings were overall relatively topically similar. • Different translation systems could also impact the reliability and effectiveness of any NLP-based approach, especially if the approach expects English data (like our approach does) and data from other languages has to be automatically translated. • Does including more metadata improve retrieval or verification performance? How should the different metadata types be included? For example, if a statement is a direct reply or a "quote tweet" of the original tweet containing the claim, it is intuitive that this type of metadata would signal increased relevance. • Multi-modality: tweets don't only contain text content, but also sometimes images and video data. Does adding this additional information to the tweet content, for example via transcription or use of multimodal capabilities of modern LLMs, improve retrieval or verification performance?</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Visualization of our system for verifying a rumor. Numbers 1-4 indicate components with multiple configuration options.</figDesc><graphic coords="3,72.00,65.61,451.26,202.83" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>Differences in average verification performance score over all configurations, for each feature. The positive difference represents the average score increase when value option 1 is used over value option 2.</figDesc><table><row><cell cols="5">Feature tested Value option 1 Value option 2 Macro-F1 Difference Strict-Macro-F1 Difference</cell></row><row><cell>Verification</cell><cell>OPENAI</cell><cell>LLAMA</cell><cell>+0.1911</cell><cell>+0.2066</cell></row><row><cell>Retrieval</cell><cell>PyTerrier</cell><cell>Embeddings</cell><cell>+0.0239</cell><cell>+0.0132</cell></row><row><cell cols="2">Preprocessing False</cell><cell>True</cell><cell>+0.0139</cell><cell>+0.0117</cell></row><row><cell cols="2">External Data False</cell><cell>True</cell><cell>+0.0066</cell><cell>+0.0078</cell></row><row><cell>Normalize</cell><cell>False</cell><cell>True</cell><cell>+0.0300</cell><cell>+0.0306</cell></row><row><cell>Scale</cell><cell>False</cell><cell>True</cell><cell>+0.0635</cell><cell>+0.0637</cell></row><row><cell>Ignore NEI</cell><cell>True</cell><cell>False</cell><cell>+0.0709</cell><cell>+0.0708</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://pyterrier.readthedocs.io/en/latest/terrier-retrieval.html</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://huggingface.co/facebook/bart-large-mnli</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">https://platform.openai.com/docs/api-reference/assistants</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">https://medium.com/@1511425435311/understanding-openais-temperature-and-top-p-parameters-in-language-models-d2066504684f</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4"><ref type="bibr" target="#b4">5</ref> https://huggingface.co/docs/hub/en/models-inference</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_5">https://platform.openai.com/docs/overview</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_6">https://huggingface.co/docs/api-inference/index</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_7">https://github.com/LuisKolb/clef-2024-authority/blob/main/clef/pipeline/eval_experiment_large.ipynb</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_8">checkthat.gitlab.io/clef2024/task5</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_9">https://platform.openai.com/docs/changelog</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="11" xml:id="foot_10">https://developer.x.com/en/docs/twitter-api/getting-started/about-twitter-api</note>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A. Online Resources</head><p>The github repository can be found at github.com/LuisKolb/clef-2024-authority. The repository includes all the different components we used for our experiments, and the scripts used to produce our results for the CheckThat! Task 5 submission.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>B. Glossary</head><p>In this paper, we use some specific words to describe specific concepts:</p><p>• "claim": the individual text snippet/sentence(s) that is to be verified (using authority sources)</p><p>• "rumor": used interchangeably with claim (in the dataset, every rumor consists of a claim and several statements, and has a "rumor_id") • "statement": a social media post, in this context posted by an authority accounts • "evidence": a statement relevant to a specific claim • "authority": typically official government social media accounts, but also sometimes the individual person a claim is about, and whose social media posts can be used to verify that claim</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>B.1. Verification Experiment Results and Tables</head><p>Configurations with the same score are assigned the same rank, as they produced the same results. Some column names are abbreviated for layout width reasons:  </p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Overview of the CLEF-2024 CheckThat! Lab: Check-worthiness, subjectivity, persuasion, roles, authorities and adversarial robustness</title>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Alam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Struß</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Chakraborty</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Elsayed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Przybyła</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Caselli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Haouari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Piskorski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Ruggeri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Suwaileh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fifteenth International Conference of the CLEF Association</title>
				<editor>
			<persName><forename type="first">L</forename><surname>Goeuriot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Mulhem</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Quénot</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><surname>Schwab</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">L</forename><surname>Soulier</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><forename type="middle">M</forename><surname>Di Nunzio</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>García Seco De Herrera</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<meeting><address><addrLine>CLEF</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2024">2024. 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Overview of the CLEF-2022 CheckThat! Lab Task 2 on Detecting Previously Fact-Checked Claims</title>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Mubarak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Babulkov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Working Notes of CLEF 2022 -Conference and Labs of the Evaluation Forum</title>
				<meeting>the Working Notes of CLEF 2022 -Conference and Labs of the Evaluation Forum<address><addrLine>Bologna, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Community-Based Fact-Checking on Twitter&apos;s Birdwatch Platform</title>
		<author>
			<persName><forename type="first">N</forename><surname>Pröllochs</surname></persName>
		</author>
		<idno type="DOI">10.1609/icwsm.v16i1.19335</idno>
		<ptr target="https://ojs.aaai.org/index.php/ICWSM/article/view/19335.doi:10.1609/icwsm.v16i1.19335" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International AAAI Conference on Web and Social Media</title>
				<meeting>the International AAAI Conference on Web and Social Media</meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="page" from="794" to="805" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Overview of the CLEF-2024 CheckThat! Lab Task 5 on Rumor Verification using Evidence from Authorities</title>
		<author>
			<persName><forename type="first">F</forename><surname>Haouari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Elsayed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Suwaileh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2024 -Conference and Labs of the Evaluation Forum, CLEF 2024</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Faggioli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Galuščáková</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>García Seco De Herrera</surname></persName>
		</editor>
		<meeting><address><addrLine>Grenoble, France</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Learning to Rank for Information Retrieval</title>
		<author>
			<persName><forename type="first">T.-Y</forename><surname>Liu</surname></persName>
		</author>
		<idno type="DOI">10.1561/1500000016</idno>
		<ptr target="https://www.nowpublishers.com/article/Details/INR-016.doi:10.1561/1500000016" />
	</analytic>
	<monogr>
		<title level="j">Foundations and Trends® in Information Retrieval</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="225" to="331" />
			<date type="published" when="2009">2009</date>
			<publisher>Now Publishers, Inc</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">The CLEF-2024 CheckThat! Lab: Check-Worthiness, Subjectivity, Persuasion, Roles, Authorities, and Adversarial Robustness</title>
		<author>
			<persName><forename type="first">A</forename><surname>Barrón-Cedeño</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Alam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Chakraborty</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Elsayed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Nakov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Przybyła</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Struß</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Haouari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hasanain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Ruggeri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Suwaileh</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-031-56069-9_62</idno>
	</analytic>
	<monogr>
		<title level="m">Advances in Information Retrieval</title>
				<editor>
			<persName><forename type="first">N</forename><surname>Goharian</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Tonellotto</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>He</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Lipani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Mcdonald</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Macdonald</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">I</forename><surname>Ounis</surname></persName>
		</editor>
		<meeting><address><addrLine>Nature Switzerland; Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2024">2024</date>
			<biblScope unit="page" from="449" to="458" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Are authorities denying or supporting? Detecting stance of authorities towards rumors in Twitter</title>
		<author>
			<persName><forename type="first">F</forename><surname>Haouari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Elsayed</surname></persName>
		</author>
		<idno type="DOI">10.1007/s13278-023-01189-3</idno>
		<ptr target="https://doi.org/10.1007/s13278-023-01189-3.doi:10.1007/s13278-023-01189-3" />
	</analytic>
	<monogr>
		<title level="j">Social Network Analysis and Mining</title>
		<imprint>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="page">34</biblScope>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">BART: Denoising Sequence-to-Sequence Pre-training for Natural Language Generation</title>
		<author>
			<persName><forename type="first">M</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ghazvininejad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mohamed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Levy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Stoyanov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.1910.13461</idno>
		<idno type="arXiv">arXiv:1910.13461</idno>
		<ptr target="http://arxiv.org/abs/1910.13461.doi:10.48550/arXiv.1910.13461" />
	</analytic>
	<monogr>
		<title level="m">Translation, and Comprehension</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note>cs, stat</note>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">Natural Language Inference</title>
		<author>
			<persName><forename type="first">B</forename><surname>Maccartney</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">U C S</forename><surname>Department</surname></persName>
		</author>
		<ptr target="https://books.google.at/books?id=F55EAQAAIAAJ" />
		<imprint>
			<date type="published" when="2009">2009</date>
		</imprint>
		<respStmt>
			<orgName>Stanford University</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">J</forename><surname>Openai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Achiam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Adler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Ahmad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">L</forename><surname>Akkaya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Aleman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Almeida</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Altenschmidt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Altman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Anadkat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Avila</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Babuschkin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Balaji</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Balcom</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Baltescu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bavarian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Belgum</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bello</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Berdine</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bernadett-Shapiro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Berner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Bogdonoff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Boiko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A.-L</forename><surname>Boyd</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Brakman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Brockman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Brooks</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Brundage</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Button</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Cai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Campbell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Cann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Carey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Carlson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Carmichael</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Chan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Chantzis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Chess</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Cho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">W</forename><surname>Chu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Chung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Cummings</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Currier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Dai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Decareaux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Degry</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Deutsch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Deville</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dhar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Dohan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Dowling</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Dunning</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ecoffet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Eleti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Eloundou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Farhi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Fedus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">P</forename><surname>Felix</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Fishman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Forte</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Fulford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Georges</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Gibson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Goel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Gogineni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Goh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gontijo-Lopes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gordon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Grafstein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Gray</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Greene</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">S</forename><surname>Gross</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Gu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Guo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hallacy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Han</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Harris</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Heaton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Heidecke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Hickey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Hickey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Hoeschele</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Houghton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Hsu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Huizinga</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Jain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Jain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Jang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Jin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Jin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Jomoto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Jonn</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Jun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Kaftan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kaiser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Kamali</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">S</forename><surname>Kanitscheider</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Keskar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Khan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">W</forename><surname>Kilpatrick</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">H</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kirchner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kiros</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Knight</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kokotajlo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kondrich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Konstantinidis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Kosic</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Krueger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Kuo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Lampe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Leike</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Leung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">M</forename><surname>Levy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Litwin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Lopez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Lowe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lue</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Makanju</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Malfacini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Manning</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Markov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Markovski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Martin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mayer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Mayne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Mcgrew</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Mckinney</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Mcleavey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Mcmillan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Mcneil</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Medina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Mehta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Menick</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Metz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Mishchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Mishkin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Monaco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Morikawa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Mossing</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Murati</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Murk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mély</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Nair</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Nakano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Nayak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Neelakantan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Ngo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Noh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Ouyang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>O'keefe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Pachocki</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Paino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Palermo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Pantuliano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Parascandolo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Parish</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Parparita</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Passos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Pavlov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Peng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">D A B</forename><surname>Perelman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Peres</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">P D O</forename><surname>Petrov</surname></persName>
		</author>
		<author>
			<persName><surname>Pinto</surname></persName>
		</author>
		<author>
			<persName><surname>Michael</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Pokorny</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">H</forename><surname>Pokrass</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Pong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Powell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Power</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Power</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Proehl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Puri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Radford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rae</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Ramesh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Raymond</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Real</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Rimbach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Ross</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Rotsted</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Roussez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ryder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Saltarelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sanders</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Santurkar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Sastry</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Schmidt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Schnurr</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Schulman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Selsam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Sheppard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sherbakov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Shieh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Shoker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Shyam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Sidor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sigler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Simens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Sitkin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Slama</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Sohl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Sokolowsky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">P</forename><surname>Staudacher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Such</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Summers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Sutskever</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">B</forename><surname>Tezak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Thompson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Tillet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Tootoonchian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Tseng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Tuggle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Turley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">F C</forename><surname>Tworek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Uribe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Vallone</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Vijayvergiya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Voss</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">J</forename><surname>Wainwright</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ward</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">J</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Weinmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Welihinda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Welinder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Weng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Weng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Wiethoff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Willner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Winter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wolrich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Wong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Workman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Xiao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Yoo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Yuan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Zaremba</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Zellers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhuang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Zhuk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zoph</surname></persName>
		</author>
		<author>
			<persName><surname>Kondraciuk</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.2303.08774</idno>
		<idno type="arXiv">arXiv:2303.08774</idno>
		<ptr target="http://arxiv.org/abs/2303.08774.doi:10.48550/arXiv.2303.08774" />
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
	<note type="report_type">GPT-4 Technical Report</note>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">L</forename><surname>Smars</surname></persName>
		</author>
		<ptr target="https://www.digitalnewsreport.org/publications/2018/factsheet-understanding-promise-limits-automated-fact-checking/" />
		<title level="m">FACTSHEET: Understanding the Promise and Limits of Automated Fact-Checking</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">The Influence of Media Trust and Normative Role Expectations on the Credibility of Fact Checkers</title>
		<author>
			<persName><forename type="first">F</forename><surname>Primig</surname></persName>
		</author>
		<idno type="DOI">10.1080/17512786.2022.2080102</idno>
		<idno>doi:10.1080/17512786.2022.2080102</idno>
		<ptr target="https://doi.org/10.1080/17512786.2022.2080102" />
	</analytic>
	<monogr>
		<title level="j">Journalism Practice</title>
		<imprint>
			<biblScope unit="volume">18</biblScope>
			<biblScope unit="page" from="1137" to="1157" />
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Who can verify this? Finding authorities for rumor verification in Twitter</title>
		<author>
			<persName><forename type="first">F</forename><surname>Haouari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Elsayed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Mansour</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.ipm.2023.103366</idno>
		<ptr target="https://www.sciencedirect.com/science/article/pii/S0306457323001036.doi:10.1016/j.ipm.2023.103366" />
	</analytic>
	<monogr>
		<title level="j">Information Processing &amp; Management</title>
		<imprint>
			<biblScope unit="volume">60</biblScope>
			<biblScope unit="page">103366</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
