<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Overview of Touché 2022: Argument Retrieval</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Alexander</forename><surname>Bondarenko</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Martin-Luther-Universität Halle-Wittenberg</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Maik</forename><surname>Fröbe</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Martin-Luther-Universität Halle-Wittenberg</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Johannes</forename><surname>Kiesel</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">Bauhaus-Universität Weimar</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Shahbaz</forename><surname>Syed</surname></persName>
							<affiliation key="aff2">
								<orgName type="institution">Leipzig University</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Timon</forename><surname>Gurcke</surname></persName>
							<affiliation key="aff3">
								<orgName type="institution">University</orgName>
								<address>
									<settlement>Paderborn</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Meriem</forename><surname>Beloucif</surname></persName>
							<affiliation key="aff4">
								<orgName type="institution">Uppsala University</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Alexander</forename><surname>Panchenko</surname></persName>
							<affiliation key="aff5">
								<orgName type="institution">Skolkovo Institute of Science and Technology</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Chris</forename><surname>Biemann</surname></persName>
							<affiliation key="aff6">
								<orgName type="institution">Universität Hamburg</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Benno</forename><surname>Stein</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">Bauhaus-Universität Weimar</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Henning</forename><surname>Wachsmuth</surname></persName>
							<affiliation key="aff3">
								<orgName type="institution">University</orgName>
								<address>
									<settlement>Paderborn</settlement>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Martin</forename><surname>Potthast</surname></persName>
							<affiliation key="aff2">
								<orgName type="institution">Leipzig University</orgName>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Matthias</forename><surname>Hagen</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Martin-Luther-Universität Halle-Wittenberg</orgName>
							</affiliation>
						</author>
						<title level="a" type="main">Overview of Touché 2022: Argument Retrieval</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">6263738CDF733D87F3F52FE41AFA597C</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T03:30+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Argument retrieval</term>
					<term>Controversial questions</term>
					<term>Comparative questions</term>
					<term>Image retrieval</term>
					<term>Shared task</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper is a report on the third year of the Touché lab on argument retrieval hosted at CLEF 2022. With the goal of supporting and promoting the research and development of new technologies for argument mining and argument analysis, we have organized three shared tasks: (a) argument retrieval for controversial topics, where the task is to find sentences that reflect the gist of arguments from online debates, (b) argument retrieval for comparative issues, where the task is to find argumentative passages from web documents that help in making a comparative decision, and (c) image retrieval for arguments, where the task is to find images that show support for or opposition to a particular stance.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Decision-making and opinion-forming are everyday tasks, often involving weighing pro and con arguments for or against different options. Considering the many arguments on almost any topic on the web, in principle anyone can come to an informed decision or opinion with the help of a search engine. However, large parts of the easily accessible arguments on the web are of low quality. They may contain incoherent logic, fail to substantiate a claim, or use inappropriate language. These arguments should not appear at the top of search results-regardless of whether a query is about socially important issues or "only" personal choices. Challenges arising from this observation range from evaluating the relevance of an argument to a query and assessing how well an implied stance is justified, to identifying the gist of an argument, to finding images that illustrate a particular stance. Commercial web search engines do not sufficiently address these challenges-a gap we aim to fill with the Touché labs.</p><p>Following the two successful Touché labs on argumentation at CLEF 2020 and 2021 <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b2">3]</ref>, our third lab edition again brought together researchers from the fields of information retrieval and natural language processing who study argumentation. At Touché 2022, we have organized the following three shared tasks, the last of which is a completely new addition: 1. Argumentative sentence retrieval from a focused collection (crawled from debate portals)</p><p>to support conversations about controversial topics. 2. Argument retrieval from a large collection of text passages to support answering comparative questions in personal decision making. 3. Argumentative image retrieval to support the illustration of arguments and getting an overview of the public opinion on controversial topics.</p><p>Touché follows the traditional TREC methodology: documents and topics are provided to participants, who then submit their results (up to five runs) for each topic to be assessed by human assessors. While the first two Touché editions focused on full argument and document retrieval, the third edition focused on more fine-grained retrieval units. The three shared tasks investigated whether argument retrieval can more directly support decision making and opinion formation by extraction of the gist of documents, classification of their stance on an issue as pro or con, and retrieval of images that support or oppose a particular stance.</p><p>The teams that participated in the third Touché lab were able to use the topics and assessments (relevance and quality of arguments) from the previous lab editions to train and optimize their approaches. In addition to traditional retrieval models such as BM25 <ref type="bibr" target="#b3">[4]</ref>, re-ranking approaches such as the recent transformer-based models T5 <ref type="bibr" target="#b4">[5]</ref> and T0 <ref type="bibr" target="#b5">[6]</ref> have been applied with the goal of combining topical relevance with "argumentativeness," argument quality, or stance. They are an essential part of the most effective approaches of all three Touché tasks, confirming the general trend in information retrieval and natural language processing that pre-trained transformers achieve good effectiveness <ref type="bibr" target="#b6">[7]</ref> (cf. . The most effective approach submitted to Task 1 re-ranks the DirichletLM model's search results by first using a BERT-based classifier <ref type="bibr" target="#b7">[8]</ref> to decide on the argumentativeness of retrieved sentence pairs (i.e., whether they are premises or assertions), then estimating their coherence using the cosine similarity of their BERT embeddings. For Task 2, in terms of relevance, a TCT-ColBERT ranker <ref type="bibr" target="#b8">[9]</ref> and, in terms of quality, a combination of query-dependent BM25F scores <ref type="bibr" target="#b9">[10]</ref> and predicted argument quality were most effective. The most effective approach for Task 3 (across topic relevance, argumentativeness, and stance relevance) used BERT instead of a stance detection model to detect the sentiment of texts from web pages and texts in images and indexed both with BM25F.</p><p>Altogether, the most effective argument retrieval approaches used various strategies for query reformulation and expansion, and for re-ranking based on estimates of argument quality or "argumentativeness". Sentiment or emotion recognition was particularly useful for the argumentative image retrieval task, as well as OCR to retrieve image text for analysis.</p><p>The corpora, topics, and judgments created at Touché are freely available to the research community and can be found on the lab's website. <ref type="foot" target="#foot_0">1</ref> Parts of the data are also already available via the BEIR <ref type="bibr" target="#b10">[11]</ref> and ir_datasets <ref type="bibr" target="#b11">[12]</ref> resources.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head><p>Queries in argument retrieval often are phrases that describe a controversial topic, questions that ask to compare two options, or even complete claims or short arguments <ref type="bibr" target="#b12">[13]</ref>. In the third edition of the Touché lab, we address the first two query types in three different shared tasks on argument retrieval in general, on comparative scenarios, and on image retrieval. Here, we briefly summarize the related work for all three tasks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Argument Retrieval</head><p>The goal of argument retrieval is to find arguments that help when making a decision, when forming an opinion, or when trying to convince (or persuade) someone of a specific point of view. An argument is usually modeled as a conclusion with one or more supporting or attacking premises <ref type="bibr" target="#b13">[14]</ref>. While a conclusion is a statement that can be accepted or rejected, a premise is a more grounded statement (e.g., statistical evidence or a referenced quote).</p><p>Adding argument retrieval components to a search engine poses challenges like identifying argumentative queries <ref type="bibr" target="#b14">[15]</ref>, mining arguments from documents, or assessing an argument's relevance and quality <ref type="bibr" target="#b13">[14]</ref>. Different paradigms have been proposed for actual argument retrieval that perform argument mining and ranking in different order <ref type="bibr" target="#b15">[16]</ref>. For instance, Wachsmuth et al. <ref type="bibr" target="#b13">[14]</ref> use distant supervision and extract and index arguments from debate portals in a "pre-processing". Their argument search engine args.me<ref type="foot" target="#foot_1">2</ref> uses BM25F <ref type="bibr" target="#b9">[10]</ref> to then only rank the extracted arguments at query time, giving more weight to conclusions than premises. Also Levy et al. <ref type="bibr" target="#b16">[17]</ref> use distant supervision to mine arguments from Wikipedia in an offline pre-processing before ranking. Following a different paradigm, Stab et al. <ref type="bibr" target="#b17">[18]</ref> retrieve documents from the Common Crawl<ref type="foot" target="#foot_2">3</ref> at query time (no prior offline argument mining) and use a topic-dependent neural network to then extract arguments from the retrieved documents. In our Touché tasks, we address both paradigms, the one of Wachsmuth et al. <ref type="bibr" target="#b13">[14]</ref> in Task 1 (retrieval from a focused collection of pre-processed arguments) and the one of Stab et al. <ref type="bibr" target="#b17">[18]</ref> in Task 2 (retrieval from some general collection with online argument mining).</p><p>Argument retrieval should take topical relevance into account but also argument quality. What makes a good argument has been studied since the time of Aristotle <ref type="bibr" target="#b18">[19]</ref>. Wachsmuth et al. <ref type="bibr" target="#b19">[20]</ref> categorize the different aspects of argument quality into a taxonomy that covers three dimensions: logic, rhetoric, and dialectic. Logic concerns the strength of the internal structure of an argument (i.e., the conclusion and the premises along with their relations) while rhetoric covers the effectiveness of an argument in persuading an audience with its conclusion. Lastly, dialectic addresses the relations of an argument to other arguments on the topic. For example, an argument attacked by many others may be rather vulnerable in a debate. Note that an argument's relevance to a query is also categorized under dialectical quality <ref type="bibr" target="#b19">[20]</ref>.</p><p>Argument relevance has been typically assessed by an argument's similarity to a given topic and by incorporating the support and attack relations to other arguments. Potthast et al. <ref type="bibr" target="#b20">[21]</ref> evaluate four standard retrieval models for ranking arguments with regard to topical relevance, logic, rhetoric, and dialectic. One of the main findings is that DirichletLM is better at ranking arguments than BM25, DPH, and TF-IDF. Gienapp et al. <ref type="bibr" target="#b21">[22]</ref> later proposed a pairwise annotation strategy that reduces the costs of crowdsourcing argument retrieval annotations by 93% (i.e., requiring the annotation of only a rather small subset of argument pairs).</p><p>As for argument ranking, several approaches exploit argument relations. For instance, Wachsmuth et al. <ref type="bibr" target="#b22">[23]</ref> connect two arguments in a graph when one uses the other's conclusion as a premise and then compute an argument's PageRank <ref type="bibr" target="#b23">[24]</ref> in this graph. In their study, taking PageRank into account improves upon baselines that only use an argument's content and internal structure (conclusion and premises) <ref type="bibr" target="#b22">[23]</ref>. Later, Dumani et al. <ref type="bibr" target="#b24">[25]</ref> used support and attack relations between clusters of premises and claims as well as between clusters of claims and a query. In an extended version, Dumani and Schenkel <ref type="bibr" target="#b25">[26]</ref> also include the quality of a premise as a probability (fraction of premises that are worse with regard to cogency, reasonableness, and effectiveness). Using a pairwise quality estimator trained on the Dagstuhl-15512 ArgQuality Corpus <ref type="bibr" target="#b26">[27]</ref>, the approach with the argument quality component was more effective on the 50 topics of Task 1 from Touché 2020 than the one without taking argument quality into account.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Retrieval for Comparisons</head><p>Comparative information needs in web search have first been addressed with basic interfaces for comparing two products entered separately in two search boxes <ref type="bibr" target="#b27">[28,</ref><ref type="bibr" target="#b28">29]</ref>. Using opinion mining approaches, comparative sentences can then be identified from product reviews in favor of or against one or the other product <ref type="bibr" target="#b29">[30,</ref><ref type="bibr" target="#b30">31,</ref><ref type="bibr" target="#b31">32]</ref>. Recently, identifying a comparison preference in a sentence (i.e., the "winning" option) has also been tackled more broadly (not just for product reviews) <ref type="bibr" target="#b32">[33,</ref><ref type="bibr" target="#b33">34]</ref> and forms the basis of the comparative argumentation machine CAM <ref type="bibr" target="#b34">[35]</ref>. Similar to the early comparison interfaces, CAM takes two objects and some comparison aspect(s) as input, retrieves comparative sentences in favor of one or the other option using BM25, and then classifies the sentences' preferences for a final merged table-like result presentation. A proper argument ranking, however, was not included in CAM. Chekalina et al. <ref type="bibr" target="#b35">[36]</ref> later extended the system to accept complete comparative questions as input and to return a natural language answer. From a comparative question, the comparison objects, aspect(s), and predicates are extracted and the system's answer is either generated directly based on transformers <ref type="bibr" target="#b7">[8]</ref> or by retrieval from an index of comparative sentences. To identify comparative questions and information needs, Bondarenko et al. <ref type="bibr" target="#b36">[37,</ref><ref type="bibr" target="#b37">38]</ref> propose a cascading ensemble of classifiers (rule-based, feature-based, and neural models). They also propose improved approaches to extract the comparison objects, aspects, and predicates from comparative questions and to detect the stance of potential answers towards the comparison objects. The respective stance dataset could also be used by the participants of our Task 2.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Image Retrieval</head><p>Images can provide contextual information and express, underline, or popularize an opinion <ref type="bibr" target="#b38">[39]</ref>, thereby taking the form of subjective statements <ref type="bibr" target="#b39">[40]</ref>. While some images can be complete arguments (i.e., expressing both, a premise and a conclusion) <ref type="bibr" target="#b40">[41,</ref><ref type="bibr" target="#b41">42]</ref> others provide contextual information only and have to be combined with a textual conclusion to form an argument. A recent SemEval task distinguished a total of 22 persuasion techniques in memes alone <ref type="bibr" target="#b42">[43]</ref>.</p><p>Moreover, argument quality dimensions like acceptability, credibility, emotional appeal, and sufficiency <ref type="bibr" target="#b26">[27]</ref> all also apply to arguments that include images.</p><p>Pre-dated only by approaches relying on metadata and similarity measures <ref type="bibr" target="#b43">[44]</ref>, the actual content of images or videos has been analyzed and used for keyword-based image search for decades <ref type="bibr" target="#b44">[45]</ref>. In a recent survey, Latif et al. <ref type="bibr" target="#b45">[46]</ref> categorize image features into color, texture, shape, and spatial features but commercial search engines also index text found in images, surrounding text, alternative texts displayed when an image is unavailable, and the image URLs <ref type="bibr" target="#b46">[47,</ref><ref type="bibr" target="#b47">48]</ref>. As for the retrieval of argumentative images, a closely related concept is "emotional images", which is based on image features like color and composition <ref type="bibr" target="#b48">[49,</ref><ref type="bibr" target="#b49">50]</ref>. Since argumentation often goes hand in hand with emotions, emotional features may also be promising for retrieving images for arguments, a relatively new task recently proposed by Kiesel et al. <ref type="bibr" target="#b50">[51]</ref> and now forming Task 3 of the Touché 2022 lab.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Lab Overview and Statistics</head><p>For the third edition of the Touché lab, we received 58 registrations, doubling the number from the previous year (29 registrations in 2021). Among the teams, 27 registered for more than one task, 17 registered particularly for Task 1, 10 for Task 2, and 4 for Task 3 (the new task this year). The majority of registrations came from Germany and Italy (13 each), followed by India <ref type="bibr" target="#b11">(12)</ref>, the United States (3), the Netherlands, France, Switzerland, Bangladesh (2 each), Pakistan, Portugal, United Kingdom, Indonesia, China, Russian Federation, Bulgaria, Nigeria, and Lebanon (1 each). Aligned with the lab's fencing-related title, the registered teams selected a real or fictional fencer or swordsman character (e.g., D'Artagnan) as their team name.</p><p>From the 58 registered teams, 23 actively participated in the tasks and submitted results<ref type="foot" target="#foot_3">4</ref> (27 teams submitted in 2021 and 17 teams in 2020). Using the setup of the previous Touché editions, we encouraged the teams to deploy their software in TIRA <ref type="bibr" target="#b51">[52]</ref> for a better reproducibility of the developed approaches. The TIRA integrated research architecture is cloud-based evaluation-as-a-service platform where shared task participants can deploy their software in a dedicated virtual machine to which they have full administrative access. By default, the virtual machines run the server version of Ubuntu 20.04 with one Intel Xeon E5-2620 CPU, 4 GB RAM, 16 GB HDD, and the latest versions of often-used software packages pre-installed (e.g., Docker and Python). If needed, we tried to customize the resources as per a team's requirements. Providing GPUs was not possible, though.</p><p>For teams that did not deploy their software in TIRA, we allowed run submissions similar to many TREC tracks. In case they preferred software submissions, the teams created their run using via web UI of TIRA by remote-executing their software inside their virtual machine. The software is fully installed in the virtual machine, and at execution time the virtual machine is shut down, disconnected from the internet, powered on again in a sandbox mode, and the test datasets for the respective tasks are mounted. Interrupting the internet connection ensures that the participants' software works without external web services that may disappear or become incompatible, which could reduce reproducibility (i.e., downloading additional external code or models during the execution is not possible). We offered support in case of problems during deployment and then archived the virtual machines that the participants used for their submissions. The respective systems can thus be re-evaluated or also applied to new datasets with the same input format.</p><p>Overall, 9 of the 23 teams submitted traditional run instead of deploying their software in TIRA. Per team, we allowed 5 runs and the run needed to follow the standard TREC format. <ref type="foot" target="#foot_4">5</ref>We checked the validity of each submitted run and asked participants to rerun their software or resubmit their files in case of problems while also offering support in such cases. In total, 84 runs were submitted-at least one from each team.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Task 1: Argument Retrieval for Controversial Questions</head><p>The goal of the Touché 2022 lab's first task was to support individuals who search for opinions and arguments on socially important controversial topics like "Are social networking sites good for our society?". Such scenarios benefit from obtaining the gists of various web resources that briefly summarize different stances (pro or con) on controversial topics. The task we considered in this regard followed the idea of extractive argument summarization <ref type="bibr" target="#b52">[53]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Task Definition and Data</head><p>Task. Given a controversial topic and a collection of arguments, the task was to retrieve sentence pairs that represent the gist of their corresponding arguments (e.g., the main claim and a supporting premise). Sentences in such a pair may not contradict each other and ideally build upon each other in a logical manner comprising a coherent text.</p><p>Topics. We used 50 controversial topics from the previous iterations of Touché. Each topic is formulated as a question that the user might pose as a query to the search engine, accompanied by a description summarizing the information need and the search scenario, along with a narrative to guide assessors in recognizing relevant results (see Table <ref type="table" target="#tab_0">1</ref>).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Document collection.</head><p>The document collection for Task 1 was based on the args.me corpus <ref type="bibr" target="#b15">[16]</ref> which contains about 400,000 structured arguments (crawled from the online debate portals debatewise.org, idebate.org, debatepedia.org, and debate.org). It is freely available for download <ref type="foot" target="#foot_5">6</ref> and can also be accessed through the args.me API. <ref type="foot" target="#foot_6">7</ref> To account for this year's changes in the task definition (the focus on gists), we prepared a pre-processed version of the corpus. Preprocessing steps included sentence splitting and removing premises and conclusions shorter than two words, resulting in 5,690,642 unique sentences with 64,633 claims and 5,626,509 premises. Description Democracy may be in the process of being disrupted by social media, with the potential creation of individual filter bubbles. So a user wonders if social networking sites should be allowed, regulated, or even banned.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Narrative</head><p>Highly relevant arguments discuss social networking in general or particular networking sites, and its/their positive or negative effects on society. Relevant arguments discuss how social networking affects people, without explicit reference to society.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Evaluation Setup</head><p>Participants submitted their rankings as traditional TREC-style runs where document IDs are sorted by descending relevance score for each search topic (i.e., the most relevant argument occurs at Rank 1). Given the large number of runs and the possibility of retrieving up to 1000 documents (in our case, these are sentence pairs) per topic in a run, using TrecTools <ref type="bibr" target="#b53">[54]</ref>, we created the pools using a top-5 pooling strategy, resulting in 6,930 unique sentence pairs for manual assessment of relevance, quality (argumentativeness), and textual coherence. Relevance was judged by our volunteer assessors on a three-point scale: 0 (not relevant), 1 (relevant), and 2 (highly relevant). For quality, annotators assessed whether a retrieved pair of sentences are rhetorically well-written on a three-point scale: 0 (low quality/non-argumentative), 1 (average quality), and 2 (high quality). Textual coherence (if the two sentences in a pair logically build upon each other) was also judged on a three-point scale: 0 (unrelated/contradicting), 1 (average coherence), and 2 (high coherence).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Submitted Approaches and Evaluation Results</head><p>This year's approaches included standard retrieval models such as TF-IDF, BM25, DirichletLM, and DPH. Participants also used third-party toolkits, such as the Project Debater API <ref type="bibr" target="#b54">[55]</ref> (for stance and evidence detection in arguments), Apache OpenNLP<ref type="foot" target="#foot_7">8</ref> (for language detection), and BERT-based classifiers proposed by Reimers et al. <ref type="bibr" target="#b55">[56]</ref> trained on the Webis Argument Quality Corpus <ref type="bibr" target="#b21">[22]</ref> and the IBM Rank 30K dataset <ref type="bibr" target="#b56">[57]</ref> for argument quality detection. Additionally, semantic similarity of word and sentence embeddings based on doc2vec <ref type="bibr" target="#b57">[58]</ref>, Spacy embeddings <ref type="bibr" target="#b58">[59]</ref>, and SBERT <ref type="bibr" target="#b59">[60]</ref> have been employed for retrieving coherent sentence pairs as required by the task definition. One team leveraged the text generation capabilities of GPT-2 <ref type="bibr" target="#b60">[61]</ref> to find subsequent sentences while another team similarly used the next sentence prediction (NSP) of BERT <ref type="bibr" target="#b7">[8]</ref> for this. These toolkits augmented the document preprocessing and re-ranking of the retrieved results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 2</head><p>Results of Task 1 (Argument Retrieval for Controversial Questions). Shown are the scores of a teams' best run for the three dimensions relevance, quality, and coherence of the retrieved sentence pairs with along a run's rank (results of all submitted runs in Tables <ref type="table">6-8</ref>). The teams are ordered alphabetically; baseline Swordsman emphasized. A † indicates statistically significant differences to the baseline (paired Student's 𝑡-test, 𝑝 = 0.05, Bonferroni-correction). We used nDCG@5 to evaluate of relevance, quality, and coherence. Table <ref type="table">2</ref> shows the results of the best run per team. On all the evaluated dimensions at least eight out of ten teams managed to beat the provided baseline. Similar to previous years' results, quality is best covered by the approaches followed by relevance and the newly added coherence dimension.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Team</head><p>Summarizing the results, for relevance, Team Porthos <ref type="bibr" target="#b61">[62]</ref> achieved the highest rank followed by Daario Naharis <ref type="bibr" target="#b62">[63]</ref> with nDCG@5 scores of 0.742 and 0.683, respectively. For the quality and coherence dimensions Daario Naharis obtained the highest scores (0.913 and 0.458) followed by Porthos (0.873 and+0.429). We believe that the two-stage re-ranking employed by Daario Naharis improved coherence and quality in comparison to the other approaches. They first ensured that retrieved pairs were relevant to their context in the argument alongside the topic which preserved high-quality arguments. Then, a second re-ranking based on stance to determine the final pairing of the retrieved sentences boosted coherence. Below, we briefly describe our baseline and summarize the submitted approaches.</p><p>Our baseline Swordsman employed a graph-based approach that ranks arguments' sentences by their centrality in the corresponding argument graph as proposed by Alshomary et al. <ref type="bibr" target="#b52">[53]</ref>. The top two sentences per argument are used as the their gist. We retrieved 1000 pairs per topic.</p><p>Bruce Banner <ref type="bibr" target="#b63">[64]</ref> employed the BM25 retrieval model implemented in the Pyserini toolkit <ref type="bibr" target="#b64">[65]</ref> with its default parameters (𝑘 1 = 1.2 and 𝑏 = 0.68). For each argument, they indexed all possible sentence pairs. To speed up computation on such a large collection of sentence pairs, they specifically opted for the sparse representations in Pyserini that produce smaller indexes compared to the dense retrieval variants. Two query variants were used: original query (topic title) and an expanded query (narrative and description appended). Likewise, two variants of the sentence pairs were indexed: original pair and pair with the topic of a debate appended. They retrieved 1000 documents per query and did not apply any re-ranking. <ref type="bibr" target="#b65">[66]</ref> also employed sparse retrieval together with text preprocessing and query expansion. For retrieval, they used two retrieval models from Lucene: BM25 <ref type="bibr" target="#b9">[10]</ref> (𝑘 1 = 1.2 and 𝑏 = 0.75) and DirichletLM (𝜇 = 2000). For preprocessing, they experimented with both Porter <ref type="bibr" target="#b66">[67]</ref> and Krovetz <ref type="bibr" target="#b67">[68]</ref> stemmers. Additionally, they filtered both character and word ngrams (referred to as shingles) and used two stop word lists (SMART System <ref type="bibr" target="#b68">[69]</ref>, Glasgow IR. <ref type="foot" target="#foot_8">9</ref> ) Query expansion was done using synonyms from WordNet <ref type="bibr" target="#b69">[70]</ref> and word2vec <ref type="bibr" target="#b70">[71]</ref>. Evaluation on the previous year's relevance judgments showed that a combination of the DirichletLM retrieval model, the Krovitz stemmer, and the Glasgow IR stop word list improved performance compared to their respective counterparts.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>D'Artagnan</head><p>Daario Naharis <ref type="bibr" target="#b62">[63]</ref> developed a standard Lucene-based document retrieval system using the TF-IDF model. Additionally, they introduced a new measure called ICoefficient for scoring the discriminant power of a term. This complements the standard TF-IDF weighting by additionally considering the number of documents that contain at least one occurrence of a given term. We refer readers to Bahrami et al. <ref type="bibr" target="#b62">[63]</ref> for the mathematical formulation of the ICoefficient. For preprocessing, they created two custom stop lists, each composed of the 100 most frequent terms in the indexed collections of the argument contexts and individual arguments from the provided corpus. Document re-ranking was performed based on stance and evidence detection using the Project Debater API <ref type="bibr" target="#b54">[55]</ref>.</p><p>Gamora <ref type="bibr" target="#b71">[72]</ref> developed Lucene-based approaches using deduplication and contextual featureenriched indexing, adding the topic of a debate and the stance on the topic, to obtain documentlevel relevance and quality scores, following the approaches used in previous Touché editions <ref type="bibr" target="#b2">[3]</ref>. To find relevant sentence pairs rather than relevant documents, these results were used to limit the number of documents by creating a new index for only the sentences of relevant documents (double indexing) or creating all possible sentence combinations and ranking them based on a weighted average of the argument quality (estimated using an SVM classifier) of the pair and its source document. BM25 <ref type="bibr" target="#b64">[65]</ref> (𝑘 1 = 1.2 and 𝑏 = 0.75) and DirichletLM (𝜇 = 2000) were used for document similarity and SBERT <ref type="bibr" target="#b59">[60]</ref> and TF-IDF for sentence similarity. The best approach is based on double indexing and a combination of a manual query reduction in which only the 2-6 main words of the query were kept, query boosting, query decorators, query expansion with respect to important keywords (GloVE <ref type="bibr" target="#b72">[73]</ref>) and synonyms (WordNet <ref type="bibr" target="#b69">[70]</ref>), and possessive removal, stemming (Krovetz stemmer <ref type="bibr" target="#b67">[68]</ref>) and length filtering of the sentences.</p><p>General Grevious <ref type="bibr" target="#b73">[74]</ref> used a conventional IR pipeline based on Lucene. First, documents were lowercased, tokenized and possessive words (with trailing ''s') were removed, keeping only tokens with a length between 3 and 20 characters. In addition, the team experimented with a variety of stemming approaches (S-stemmer <ref type="bibr" target="#b74">[75]</ref>, Krovetz stemmer <ref type="bibr" target="#b67">[68]</ref>, Porter stemmer <ref type="bibr" target="#b66">[67]</ref>, no stemming) and stop word lists (Core NLP <ref type="bibr" target="#b75">[76]</ref>, CountWordsFree <ref type="bibr" target="#b64">[65]</ref>, EBSCO, <ref type="foot" target="#foot_9">10</ref> GoogleStop, <ref type="foot" target="#foot_10">11</ref>and Ranks. <ref type="foot" target="#foot_11">12</ref> ) To retrieve documents, BM25 <ref type="bibr" target="#b64">[65]</ref> (𝑘 1 = 1.2 and 𝑏 = 0.75) and DirichletLM (𝜇 ∈ {1700, 1800}) were used together with query boosting, by assigning weights to the used inputs (argument, conclusion, debate title, and argument title), and query expansion, by finding keywords (Rapid Automatic Keyword Extraction (RAKE) <ref type="bibr" target="#b76">[77]</ref>) and synonyms (Datamuse <ref type="foot" target="#foot_12">13</ref> ). This retrieval step was done once for the documents and once for all the potential sentence pairs within these retrieved documents to obtain a ranking of sentence pairs. Finally, sentiment analysis (Vader <ref type="bibr" target="#b77">[78]</ref>) was used to boost documents that have a similar sentiment as the query, and readability analysis (Flesch-Kincaid <ref type="bibr" target="#b78">[79]</ref>) was used for re-ranking. Their best model does not include re-ranking, stemming, and stop word removal but relies solely on the combination of query expansion and the BM25 retrieval model.</p><p>Gorgon <ref type="bibr" target="#b79">[80]</ref> also used a Lucene-based IR pipeline and compared BM25 <ref type="bibr" target="#b64">[65]</ref> (𝑘 1 = 1.2 and 𝑏 = 0.75) and DirichletLM (𝜇 = 2000) similarity measures, developing four different analyzers with different preprocessing steps including lowercasing, stemming (Krovetz stemmer <ref type="bibr" target="#b67">[68]</ref>), removing possessive words (with trailing ''s') and filtering stop words (99webtools, <ref type="foot" target="#foot_13">14</ref> EBSCO). Sentence pairs were created from all combinations within a single document before indexing. The best approach is a combination of lowercasing, removing possessive words, and BM25.</p><p>Hit Girl <ref type="bibr" target="#b80">[81]</ref> proposed a two-stage retrieval pipeline that combines semantic search and re-ranking via argument quality agnostic models. Documents were embedded to vectors using Spacy <ref type="bibr" target="#b58">[59]</ref>. These were then indexed via Elasticsearch and its text similarity function used for semantic search. They experimented with three approaches for re-ranking: maximal marginal relevance <ref type="bibr" target="#b81">[82]</ref>, word mover's distance <ref type="bibr" target="#b82">[83]</ref>, and a novel method called structural distance which employs fuzzy matching between query and sentences based on POS tags. Preliminary evaluations showed that, while re-ranking improved the argument quality to varying degrees, it also affected relevance. Also, structural distance performed best for re-ranking.</p><p>Korg's <ref type="bibr" target="#b83">[84]</ref> approaches are based on the Elasticsearch implementation of DirichletLM (𝜇 = 2000) to find the best matching argumentative sentences for a query after employing lowercasing, ASCII folding, stop word filtering (manually created stop word list) and stemming (Krovetz stemmer <ref type="bibr" target="#b67">[68]</ref>). Then, either doc2vec <ref type="bibr" target="#b57">[58]</ref> or SBERT <ref type="bibr" target="#b59">[60]</ref> is trained on all sentences in the args.me corpus, which was used to find the most similar sentence pair within a document by direct comparison of the doc2vec embeddings. Alternatively, instead of directly comparing sentences, GPT-2 <ref type="bibr" target="#b60">[61]</ref> was used to generate the next sentence for a given sentence to then find the most similar sentence to the generated sentence. The best approach is based on lowercasing, ASCII folding, stop word filtering, stemming, and doc2vec's similarity calculation without GPT-2.</p><p>Pearl <ref type="bibr" target="#b84">[85]</ref> also proposed a two-stage retrieval pipeline using DirichletLM <ref type="bibr" target="#b85">[86]</ref> and DPH [87] models to retrieve argumentative sentences. For both stages, they used the PyTerrier toolkit <ref type="bibr" target="#b87">[88]</ref>. After retrieving the documents, two BERT-based argument quality models fine-tuned on the Webis Argument Quality Corpus <ref type="bibr" target="#b88">[89]</ref>, and the IBM-Rank-30k dataset <ref type="bibr" target="#b56">[57]</ref> were employed to filter non-argumentative results. The resulting prototype from the first stage was considered the baseline model. On evaluating this on a set of 35 queries taken from the provided topics, they found that the DPH model assigned high relevance to sentences even if their terms are part of a URL, or other meta data in the corpus. Moreover, it was also susceptible to homonyms and thus negatively affecting the retrieval performance. To account for this, a refined prototype was developed that combined argument quality prediction with query expansion. For query expansion, they applied the Bo1 query expansion algorithm provided by PyTerrier which weighs the terms based on divergence from randomness build on Bose-Einstein statistics <ref type="bibr" target="#b89">[90]</ref>. Specifically, the Bo1 model extracts terms from the top-ranked documents retrieved for the original query, weighs them based on their informativeness, and appends the highest-weighted terms to the original query to expand it. Finally, a custom block list consisting of commonly repeated phrases such as "my opponent claims... ", "PRO claims... ", "I accept this debate" filtered further noisy sentences, leading to improved nDCG scores.</p><p>Porthos <ref type="bibr" target="#b61">[62]</ref> used the Elasticsearch implementation of DirichletLM (with 𝜇 = 116 being the average length of sentences in the corpus) and BM25 <ref type="bibr" target="#b64">[65]</ref> (default Elasticsearch implementation with 𝑘 1 = 1.2 and 𝑏 = 0.75) or retrieval after removing sentence duplicates and filtering non-relevant sentences by removing ones with low-quality language to retain only the ones that contain at least one verb. Another filtering step is based on the argumentativeness of sentences using the support vector machine (SVM) of <ref type="bibr" target="#b21">[22]</ref> and the BERT approach of <ref type="bibr" target="#b55">[56]</ref>. In addition, sentences were stemmed, lowercased and stop words were removed. The approaches are based on a search term as a composition of single terms and Boolean queries together with Reimers et al. <ref type="bibr" target="#b55">[56]</ref> to reorder the retrieved sentences according to their argumentative quality. The sentences are paired with SBERT <ref type="bibr" target="#b59">[60]</ref> and BERT <ref type="bibr" target="#b7">[8]</ref> trained on Next Sentence Prediction (NSP). The best approach is based on DirichletLM, NSP, using the sentence classifier in preprocessing, Boolean queries with Noun Chunking for retrieval, and the BERT approach of <ref type="bibr" target="#b55">[56]</ref> for re-ranking.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Task 2: Argument Retrieval for Comparative Questions</head><p>The goal of the Touché 2022 lab's second task was to support informed decisions in "everyday" or personal comparison situations-for instance for a question like "Should I major in philosophy or psychology?". Decision making in such situations benefits from finding balanced reasons for choosing one option over the other, usually in form of opinions or arguments.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Task Definition and Data</head><p>Task. Given a collection of text passages and a comparative topic with two comparison objects, the task was to retrieve relevant argumentative passages for or against one or both objects, and to detect the passages' stances with respect to the objects.</p><p>Topics. We provided 50 topics that describe scenarios of personal decision making. Each topic has a title formulated as a comparative question, a pair of comparison objects from the title that could be used for the stance detection of the retrieved passages, a description with some background on the particular search scenario, and a narrative that served as a guideline for our assessors (cf. Table <ref type="table" target="#tab_2">3</ref> for an example). Document collection. The retrieval collection for Task 2 was a corpus of 868,655 passages extracted from ClueWeb12. <ref type="foot" target="#foot_14">15</ref> We constructed this passage corpus using all 37,248 documents from the top-100 pool of all runs submitted to Task 2 in the previous Touché editions. Using the Description A soon-to-be high-school graduate finds themself at a crossroad in their life. Based on their interests, majoring in philosophy or in psychology are the potential options and the graduate is searching for information about the differences and similarities, as well as advantages and disadvantages of majoring in either of them (e.g., with respect to career opportunities or gained skills).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Narrative</head><p>Relevant documents will overview one of the two majors in terms of career prospects or developed new skills, or they will provide a list of reasons to major in one or the other. Highly relevant documents will compare the two majors side-by-side and help to decide which should be preferred in what context. Not relevant are study program and university advertisements or general descriptions of the disciplines that do not mention benefits, advantages, or pros/cons.</p><p>TREC CAsT tools, <ref type="foot" target="#foot_15">16</ref> we split the documents at sentence boundaries into fixed-length passages of approximately 250 terms, since ranking fixed-length passages was shown to be more effective than that of variable-length passages <ref type="bibr" target="#b90">[91]</ref>. From the initial 1,286,977 passages, we removed near-duplicates with CopyCat <ref type="bibr" target="#b91">[92]</ref> to mitigate unwanted side-effects of near-duplicates on retrieval effectiveness <ref type="bibr" target="#b92">[93,</ref><ref type="bibr" target="#b93">94]</ref>, resulting in the final collection of 868,655 passages. We also provided a second version of the corpus, in which the passages were expanded with queries generated by the docT5query model <ref type="bibr" target="#b94">[95]</ref>.</p><p>To lower the bar to entry of this task, we also provided the participants with a number of previously compiled resources. These included the document-level relevance and argument quality judgments from the previous Touché editions as well as the passage-level relevance judgments from a subset of MS MARCO <ref type="bibr" target="#b95">[96]</ref> with about 40,000 comparative questions identified by an ALBERT-based <ref type="bibr" target="#b96">[97]</ref> classifier <ref type="bibr" target="#b37">[38]</ref>. Each question in MS MARCO is associated with 10 text passages (one is labeled as most relevant). To train stance detectors, an annotated dataset of 950 comparative questions and answers, extracted from Stack Exchange, was also provided <ref type="bibr" target="#b37">[38]</ref>. For the identification of claims and premises, the participants could use any own or existing argument tagging tool, such as the API <ref type="foot" target="#foot_16">17</ref> of TARGER <ref type="bibr" target="#b97">[98]</ref> hosted on our own servers.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">Evaluation Setup</head><p>Similar to Task 1, we pooled the top-5 passages from the runs, resulting in 2,107 unique passages that were manually judged. Our volunteer human assessors labeled the passages' relevance with three labels: 0 (not relevant), 1 (relevant), and 2 (highly relevant). They also assessed whether arguments are present in a passage and whether they are rhetorically well-written <ref type="bibr" target="#b26">[27]</ref> with three labels: 0 (low quality, or no arguments in a passage), 1 (average quality), and 2 (high quality). Finally, we asked the assessors to label passages with respect to a topic's comparison objects as (a) pro first object, (b) pro second object, (c) neutral (both comparison objects are equally good or bad), and (d) no stance (no stance given). In Task 2, we used nDCG@5 for the relevance and argument quality dimensions and macro-averaged F 1 for the stance detection.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.">Submitted Approaches and Evaluation Results</head><p>Seven teams submitted their results to Task 2 (25 valid runs). Interestingly, only two teams used relevance judgments from the previous Touché editions to fine-tune their models or to optimize parameters. The others either manually labeled a sample of retrieved documents themselves or relied on zero-shot approaches like the transformer-based model T0++ <ref type="bibr" target="#b5">[6]</ref>. Most teams used the standard passage collection, but two teams also used the docT5query-expanded <ref type="bibr" target="#b94">[95]</ref> collection provided by us. Overall, the main trend of this year was the usage of transformer-based models for ranking and re-ranking (e.g., ColBERT <ref type="bibr" target="#b98">[99]</ref> or monoT5 and duoT5 <ref type="bibr" target="#b99">[100]</ref>) while our baseline approach was BM25, as in the previous years.</p><p>For the optional subtask of stance detection, five of the seven teams submitted results. They either trained their own classifiers on the provided stance dataset, fine-tuned pre-trained language models, or directly used pre-trained models as zero-shot classifiers. Our baseline stance detector was a simple always-'no stance' predictor (majority class).</p><p>Table <ref type="table">4</ref> shows the results of each team's most effective runs with respect to relevance and argument quality (more detailed results for each submitted run can be found in Appendix A). For stance detection, for each team, we evaluated all passages that were part of the manual judgment pool and for which the team had predicted a stance (i.e., the stance of a passage returned at Rank 3 by some Team X (and thus part of the judgment pool) was also used in the stance evaluation of Team Y, even when the document was only on Rank 6 or lower (and thus not actually part of the pool for that run). Note that this potentially yields different numbers of passages used for the stance evaluation per team. Below, we briefly describe the teams' submitted approaches and their results (teams ordered by their relevance-wise best approach).</p><p>Captian Levi <ref type="bibr" target="#b100">[101]</ref> submitted the relevance-wise most effective run. They first retrieved 2,000 documents using Pyserini's BM25 <ref type="bibr" target="#b64">[65]</ref> (𝑘 1 = 1.2 and 𝑏 = 0.68) by combining top-1000 results for the original query (topic title) with the results for modified queries, where they used alternative strategies: (1) only removing stop words (using the NLTK <ref type="bibr" target="#b101">[102]</ref> stop word list), (2) replacing comparative adjectives with synonyms and antonyms found in WordNet <ref type="bibr" target="#b69">[70]</ref>, (3) adding extra terms using pseudo-relevance feedback, (4) using queries generated with the docT5query model <ref type="bibr" target="#b94">[95]</ref> provided by the Touché organizers. Queries and corpus were also processed by using stop words and punctuation removal and lemmatization (WordNet lemmatizer). The initially retrieved results were re-ranked using monoT5 and duoT5 <ref type="bibr" target="#b99">[100]</ref>. Additionally, TCT-ColBERT <ref type="bibr" target="#b8">[9]</ref> (a variant of ColBERT <ref type="bibr" target="#b98">[99]</ref> with knowledge distillation) was also used for initial ranking for unmodified queries (topic titles). Captain Levi submitted in total five runs that differ in the aforementioned strategies of modifying queries, initial ranking models, and final re-ranking models. Their most effective run in terms of relevance and quality was initial ranking by TCT-ColBERT. Finally, stance was detected using a RoBERTa-Large-MNLI</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 4</head><p>Results of Task 2 (Argument Retrieval for Comparative Questions). (a) Evaluation results of a team's best run according to the results' relevance. (b) Best runs according to the results' quality. (c) Stance detection results (the teams' ordering is the same as in (b)). An asterisk ( ⋆ ) indicates that the runs with the best relevance and the best quality differ for a team. The baseline BM25 ranking is shown in bold; the baseline stance detector always predicts 'no stance'. A † indicates statistically significant differences to the baseline (paired Student's 𝑡-test, 𝑝 = 0.05, Bonferroni-correction). Since stance detection results were calculated for different numbers of predictions for each team, we do not test statistical differences. Tables <ref type="table" target="#tab_0">9-11</ref> show the results for all submitted runs.</p><p>(a) Best relevance score per team , pre-trained on the Multi-Genre Natural Language Inference corpus <ref type="bibr" target="#b103">[104]</ref> without further fine-tuning in two steps: (1) detecting if the document has a stance, and then (2) for documents that were not classified as 'neutral' or 'no stance', detecting which comparison object the document favors. This stance detector achieved the highest macro-averaged F 1 score.</p><p>Aldo Nadi <ref type="bibr" target="#b104">[105]</ref> submitted the quality-wise most effective run. They re-ranked passages that were initially retrieved with BM25F <ref type="bibr" target="#b9">[10]</ref> (default Lucene implementation with 𝑘 1 = 1.2 and 𝑏 = 0.75) on two fields: text of the original passages, and passages expanded with docT5query. All texts were processed with the Porter stemmer <ref type="bibr" target="#b66">[67]</ref>, removing stop words using different lists: (a) Snowball <ref type="bibr" target="#b105">[106]</ref>, (b) a default Lucene stop word list, (c) a custom list containing the 400 most frequent terms in the retrieval collection, excluding the comparison objects. Queries (topic titles) were expanded using a relevance feedback method based on the Rocchio Algorithm <ref type="bibr" target="#b106">[107]</ref>. For the final ranking, the team experimented with two re-ranking techniques (involving up to the top-1000 documents from the initial results): (1) exploiting the argument quality estimation, i.e., they multiplied the document relevance and the quality scores, and (2) Reciprocal Ranking Fusion <ref type="bibr" target="#b107">[108]</ref>. The quality scores were predicted using the IBM Project Debater API <ref type="bibr" target="#b54">[55]</ref>. Aldo Nadi submitted five runs, which vary by different combinations of the proposed methods, e.g., using different stop word lists for pre-processing, using relevance feedback or not, using the quality-based re-ranking or fusion. The team's most effective run in terms of relevance used relevance feedback, and the most effective run in terms of quality was based on Reciprocal Ranking Fusion. The did not detect the stance.</p><p>Katana <ref type="bibr" target="#b108">[109]</ref> submitted three runs that all used different variants of ColBERT <ref type="bibr" target="#b98">[99]</ref>: (1) pretrained on MS MARCO <ref type="bibr" target="#b95">[96]</ref> by the University of Glasgow, <ref type="foot" target="#foot_17">18</ref> (2) pre-trained by Katana from scratch on MS MARCO, replacing a cosine similarity between a query and a document representation with L2 distance, and (3) the latter model fine-tuned on the relevance and quality judgments from the previous Touché editions. As queries the team used topic titles without additional processing. The team's most effective run in terms of relevance used ranking by pre-trained ColBERT, and the most effective run in terms of quality used ranking by training Col-BERT from scratch (without further fine-tuning). For stance detection, Katana used a pre-trained XGBoost-based classifier that is part of Comparative Argumentation Machine <ref type="bibr" target="#b34">[35,</ref><ref type="bibr" target="#b32">33]</ref>.</p><p>Captain Tempesta <ref type="bibr" target="#b109">[110]</ref> exploited linguistic properties of text such as a non-informative symbol frequency (hashtags, emojis, etc.), a difference between a short words' (less or equal than 4 characters) frequency and a long words' (more than 4 characters) frequency, and adjective as well as comparative adjective frequencies. Based on these properties for each document in the retrieval corpus, a quality score was computed as a weighted sum (weights were assigned manually). At query time, the relevance score of BM25 (Lucene; default: 𝑘 1 = 1.2 and 𝑏 = 0.75) was multiplied with the quality score, used as ranking criterion. Queries (topic titles) were processed by removing stop words (Lucene default list) and lowercasing query terms except for brand names, <ref type="foot" target="#foot_18">19</ref> stemming them using Lovins stemmer <ref type="bibr" target="#b110">[111]</ref>. The team's five submitted runs differ in the weights manually assigned for the different quality properties. The team's most effective run in terms of relevance used document quality estimation with linguistic properties, and the most effective run in terms of quality did not. The team did not detect stance.</p><p>Olivier Armstrong <ref type="bibr" target="#b111">[112]</ref> submitted one run. They first identified the comparison objects, aspects, and predicates in queries (topic titles) using a RoBERTa-based classifier proposed by Bondarenko et al. <ref type="bibr" target="#b37">[38]</ref>. After removing stop words, queries were expanded with synonyms of the objects, aspects, and predicates found using WordNet. Then 100 documents were retrieved using Elasticsearch's BM25 (𝑘 1 = 1.2 and 𝑏 = 0.75) as initial ranking. Using a DistilBERT-based classifier <ref type="bibr" target="#b112">[113]</ref>, fine-tuned by Alhamzeh et al. <ref type="bibr" target="#b113">[114]</ref> (a Touché 2021 participant), Olivier Armstrong identified premises and claims in the retrieved documents. For ranking, the following scores were calculated for each candidate document: (1) the arg-BM25 score returned by querying the new re-indexed corpus (only premises and claims are kept) using the unmodified queries (topic titles), (2) the argument support score, i.e., the ratio of premises and claims in the document, (3) the similarity score, i.e., the averaged cosine similarity between the original query and every premise and claim in the document represented using the SBERT embeddings <ref type="bibr" target="#b59">[60]</ref>. The final score for each candidate document was calculated as sum of the normalized individual scores. Their final ranking included 25 documents. For stance detection, the team used an LSTM-based neural network with one hidden layer that was pre-trained on the provided stance dataset.</p><p>Puss in Boots was our baseline retrieval model that used the BM25 implementation in Pyserini <ref type="bibr" target="#b64">[65]</ref> with default parameters (𝑘 1 = 0.9 and 𝑏 = 0.4) and original topic titles as queries. The baseline stance detector simply assigned 'no stance' to all documents in the ranked list.</p><p>Grimjack <ref type="bibr" target="#b114">[115]</ref> submitted five runs using query expansion and query reformulation, argument quality estimation, stance detection, and axiomatic re-ranking. For the first ranking result, the team simply retrieved 100 passages ranked with a Pyserini implementation of DirichletLM (default 𝜇 = 1000), using original, unmodified queries (topic titles). Another approach re-ranked the top-10 of the initially retrieved passages using (1) argument axioms that "prefer" documents with more premises and claims (identified with TARGER <ref type="bibr" target="#b97">[98]</ref>) or earlier occurrence of query terms in premises and claims <ref type="bibr" target="#b115">[116,</ref><ref type="bibr" target="#b116">117]</ref>, (2) newly proposed comparative axioms that "prefer" documents with more comparison objects or their earlier occurrence in premises and claims, and (3) an argument quality axiom that ranks higher documents with higher argument quality scores calculated using the IBM Project Debater API <ref type="bibr" target="#b54">[55]</ref>. For another result ranking, document positions (from the previous run) were changed based on the predicted stance, such as the 'pro first object' document was followed by the 'pro second object' followed by 'neutral' stance. The document stance was predicted using the IBM Project Debater API <ref type="bibr" target="#b54">[55]</ref>. The last two runs used T0++ <ref type="bibr" target="#b5">[6]</ref> (1) to expand queries, e.g., by combining topic titles with newly generated queries, where T0++ was prompted to generate a question given a topic's description, <ref type="bibr" target="#b1">(2)</ref> to assess the argument quality, and (3) to detect the stance in zero-shot settings. These two runs differed in whether a stance balancing was used. The team's most effective run in terms of relevance and quality used axiomatic re-ranking, and re-ranking based on the detected stance.</p><p>Asuna <ref type="bibr" target="#b117">[118]</ref> preprocessed each document (passage) in the retrieval corpus by ( <ref type="formula">1</ref>) creating a one-sentence extractive summary using LexRank <ref type="bibr" target="#b118">[119]</ref>, (2) identifying premises and claims with TARGER <ref type="bibr" target="#b97">[98]</ref>, and (3) looking up the spam score in the Waterloo Spam Rankings dataset <ref type="bibr" target="#b119">[120]</ref>. <ref type="foot" target="#foot_19">20</ref>The modified corpus was indexed, and initial retrieval of the top-40 documents was performed with the Pyserini <ref type="bibr" target="#b64">[65]</ref> implementation of BM25F (default 𝑘 1 = 0.9 and 𝑏 = 0.4) using the unmodified queries (topic titles) over the index fields with original passages, summaries, and premises and claims. Next, the queries were lemmatized and stop words were removed using the NLTK library, and expanded with the most frequent terms coming from LDA topics <ref type="bibr" target="#b120">[121]</ref> for the initially retrieved documents. The expanded queries were used to, again, retrieve the top-40 passages with BM25F. Finally, Asuna re-ranked the retrieved documents using a random forest classifier <ref type="bibr" target="#b121">[122]</ref> with the following features: BM25F score, number of times the document was retrieved for different queries (original, three extended with the LDA topics for documents, and one extended with the LDA topic for the task topic description), number of tokens in documents, number of sentences in documents, number of premises in documents, number of claims in documents, spam-score, predicted argument quality score, and predicted stance. The classifier was trained on the Touché 2020 and 2021 relevance judgments. The argument quality was predicted using DistilBERT, fine-tuned on the Webis-ArgQuality-20 corpus <ref type="bibr" target="#b88">[89]</ref>. The stance was also predicted using DistilBERT, fine-tuned on the provided stance dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Task 3: Image Retrieval for Arguments</head><p>The goal of the Touché 2022 lab's third task was to provide argumentation support through image search. The retrieval of relevant images should provide both a quick visual overview of frequent arguments on some topic, and for compelling images to support one's argumentation. The goal of the third task was thus to retrieve images that indicate an agreement or disagreement to some stance on a given topic as two separate lists similar to textual argument search.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.1.">Task Definition and Data</head><p>Task. Given a controversial topic, the task was to retrieve images (from web pages) for each stance (pro and con) that show support for that stance.</p><p>Topics. Task 3 uses the same 50 controversial topics as Task 1 (cf. Section 4). Document collection. This task's document collection stems from a focused crawl of 23,841 images and associated web pages from late 2021. For each of the 50 topics, we issued 11 queries (with different filter words like "good, " "meme, " "stats, " "reasons, " or "effects") to Google's image search and downloaded the top 100 images and associated web pages; 868 duplicate images were identified and removed using pHash <ref type="foot" target="#foot_20">21</ref> and manual checks. The dataset contains for each image: (1) the image itself in both WebP and PNG format, (2) its URL; (3) its pHash. Moreover, the dataset contains for each page: (1) its URL; (2) the Google rank of the page for each query for which the image was retrieved; (3) a WARC web archive; <ref type="foot" target="#foot_21">22</ref> (4) a DOM HTML snapshot;</p><p>(5) its complete text; (6) a screenshot; (7) meta-information of each DOM node, including the node's xPath, CSS attributes, and position on the screenshot; and (8) the xPath of the corresponding image in the DOM HTML snapshot. The full dataset is 368 GB large. <ref type="foot" target="#foot_22">23</ref> To kickstart machine learning approaches, we provided 334 relevance judgments from Kiesel et al. <ref type="bibr" target="#b50">[51]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.2.">Evaluation Setup</head><p>We employed crowdsourcing on Amazon Mechanical Turk <ref type="foot" target="#foot_23">24</ref> to evaluate the topical relevance, argumentativeness, and stance of the 6,607 image-topic pairs from all runs, employing 5 independent annotators each. Specifically, we asked for each topic for which an image was retrieved:</p><p>(1) Is the image in some manner related to the topic? (2) Do you think most people would say that, if someone shares this image without further comment, they want to show they approve of the pro-side to the topic? (3) Or do you think most people would rather say the one who shares this image does so to show they disapprove? We described each topic using the topic's title, modified as necessary to convey the description and narrative (cf. Table <ref type="table" target="#tab_0">1</ref>) and to clarify which stance is approve (pro) and disapprove (con). We then iteratively employed MACE <ref type="bibr" target="#b122">[123]</ref> to identify image-topic pairs with low annotator agreement (MACE confidence ≤ 0.55) and re-judged them ourselves, employing our judgments as check instances for another iteration of MACE. We repeated this procedure until MACE predicted the labels for all image-topic pairs from the runs with a confidence above 0.55 (re-judging 2,056 images total).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.3.">Submitted Approaches and Evaluation Results</head><p>In total, 3 teams submitted 12 runs to this task. The teams pursued quite different approaches. However, all participants employed OCR (specifically Tesseract<ref type="foot" target="#foot_24">25</ref> ) to extract image text. The</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 5</head><p>Results of Task 3 (Image Retrieval for Arguments) in terms of Precision@10 (per stance) for topic relevance, argumentativeness, and stance relevance. The table shows the best run for each team across all three measures. Results for the baseline are shown in bold. A † indicates statistically significant differences to the baseline (paired Student's 𝑡-test, 𝑝 = 0.05, Bonferroni-correction). Table <ref type="table" target="#tab_0">12</ref> shows the results for all submitted runs. teams Boromir and Jester also used the associated web page's text, but Team Jester restricted to text close to the image on the web page. Each team used sentiment or emotion features, based on image colors (Aramis), faces in the images (Jester), image text (all), and the web page text (Boromir, Jester). Boromir used the ranking information for internal evaluation. We used Precision@10 for evaluation: the ratio of relevant images among 10 retrieved images for each topic and stance. Table <ref type="table">5</ref> shows the results of each team's most effective run. For each team, the best runs were the same with respect to all three measures.</p><p>Minsc represents our baseline run, which ranks images in the same order as our original Google queries, namely of the query that includes the filter word "good" for pro and of the query that includes "anti" for con. We considered this a tough baseline, especially for on-topic relevance, as topical relatedness is similar for argumentative and "standard" web image search. However, Boromir beat this baseline-with a considerable margin for on-topic relevance.</p><p>Aramis <ref type="bibr" target="#b123">[124]</ref> focused on image features. No retrieval model was employed, but all images evaluated for each topic. They tested the use of a heuristic formula vs. fully-connected neural network classifiers for both argumentativeness and stance detection. Features were based on OCR (text length in characters, text area size, and cells in an 8×8 grid with high text density, VADER sentiment score <ref type="bibr" target="#b77">[78]</ref>), image color (average color, dominant color, and percentages of pixels with each of these color ranges as per self-defined RGB-buckets: red, green, blue, yellow, light, and dark), image category (graphic vs. photo <ref type="bibr" target="#b124">[125]</ref>; percentage of area covered by diagrams 26 ), and query-text similarity (whether the query is fully contained, the overlap for an optimal query alignment, and VADER sentiment score of words in a six-token radius around occurrences of query terms in the text). However, the query-text similarity features were not used for argumentativeness classification, as the team assumed this sub-task to be queryindependent. In our evaluation, the formula performed better than the neural approaches, which Aramis traced back to the formula being slightly better at handling off-topic images-with topical relevance not being the team's focus, they had trained and internally evaluated the network on on-topic images only. However, their worst runs still achieved a similar Precision@10 as their best one, namely 0.664 (topic; -0.037 compared to best run), 0.609 (argumentativeness; -0.025), and 0.344 (stance; -0.037). Moreover, for an evaluation that ignores the problem of topical relevance, the ratio of argumentative images among topical relevant images for their runs is between 0.904 (using both formulas) and 0.927 (using both networks), and thus very close to the baseline, which reaches a ratio of 0.932.</p><p>Boromir <ref type="bibr" target="#b125">[126]</ref> indexed both image text (boosted five-fold) and web page text (Elasticsearch BM25 with default settings, 𝑘 1 = 1.2 and 𝑏 = 0.75), using lowercasing, URL, punctuation and number removal, NLTK's WordNet lemmatization <ref type="bibr" target="#b101">[102]</ref>, removal of tokens consisting of exactly one letter, stop word removal (using the list from NLTK), and min-frequency filtering (removing tokens that appear less than three times in the text). They clustered images into 13 clusters (as determined by the elbow criterion) using 𝑘-Means and manually assigned retrieval boosts per cluster to favor more argumentative images, especially diagrams. For example, clusters with the highest boost of 5.0 were found to contain, upon manual inspection, "graphics with text (e.g., memes, quotes, twitter posts)," "graphics with round forms and text (e.g., pie charts)," "statistical graphics but with better quality [...] (e.g., bar plots, tables, line plots)," and "statistical plots (bar plots and line plots)." On the other hand, not boosted were images from clusters that were found to contain mostly photos (five clusters). They employed textual sentiment detection for stance detection, using either a dictionary (AFINN <ref type="bibr" target="#b126">[127]</ref>) or a BERT classifier. Their approach performed best and convincingly improved over the baseline. The BERT classifier improved over the dictionary-based classifier whereas image clustering was detrimental. Specifically, the image clustering seemed to introduce more off-topic images into the ranking: the same setup as the best run but using image clusters achieved a Precision@10 of 0.822 (topic; -0.056), 0.728 (argumentativeness; -0.040), and 0.411 (stance; -0.014).</p><p>Jester focused on emotion-based image retrieval via facial image recognition (using FER <ref type="foot" target="#foot_25">27</ref> ), image text, and the associated web page's text that is close to the image in the HTML source code-for which they use the text within the image's parent element. Similar to stance detection in the args.me search engine <ref type="bibr" target="#b13">[14]</ref>, they assign positive-leaning images to the pro-stance and negative-leaning images to the con-stance. For comparison, they submitted a second run without emotion features (thus plain retrieval), which achieved a lower Precision@10: 0.671 (topic; -0.025), 0.618 (argumentativeness; -0.029), and 0.336 (stance; -0.014). Thus emotion features seem helpful but insufficient when taken alone.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Conclusion</head><p>The third edition of the Touché lab at CLEF 2022 featured three shared tasks: (1) argument retrieval for controversial questions, (2) argument retrieval for comparative questions, and (3) image retrieval for arguments. Compared to previous editions, retrieval units have been changed (sentences/passages instead of full arguments/documents and images as a completely new unit) and stance detection has been included. Of 58 registered teams, 23 participated in the tasks and submitted at least one valid run. In addition to sparse retrieval and various query processing, reformulation, and expansion methods, approaches have increasingly focused on transformer models and re-ranking techniques. Not only was the quality of the documents and arguments evaluated, but also the predicted stance taken into account for the final rankings.</p><p>The most effective approaches to argument retrieval all share common characteristics. For example, most use various strategies for query reformulation and expansion, such as synonyms, relevance feedback, or generating new queries with pre-trained language models. An interesting observation is that re-ranking first-stage search results based on a quality assessment of the arguments almost always improves retrieval effectiveness. Specifically for Task 2 ( comparative questions), re-ranking based on important terms such as comparison objects and aspects or argument units in documents (premises and assertions) was successful. In task 2, stance detection was a new subtask, and some participants included a re-ranking step based on the predicted stance in their retrieval pipelines, which had some promising effects on retrieval effectiveness. However, the overall still rather low effectiveness of the approaches to stance detection leaves room for future improvements. For Task 3 (image retrieval), the recognition of sentiment and emotion and the use of OCR to analyze the text in images were particularly helpful.</p><p>We plan to continue Touché as a collaborative platform for researchers in argument retrieval. All Touché resources are freely available, including topics, manual relevance and argument quality assessments, and submitted runs from participating teams. These resources, the submission and evaluation tools, and other events such as workshops will help to further foster the community working on argument retrieval. In the future, we plan to expand the evaluation pools and to include additional dimensions of argument quality. Improving stance detection and exploiting predicted stances better not only for ranking text arguments but also for images are also interesting tasks for future work.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Example topic for Task 1: Argument Retrieval for Controversial Questions.</figDesc><table><row><cell>Number</cell><cell>34</cell></row><row><cell>Title</cell><cell>Are social networking sites good for our society?</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc>Example topic for Task 2: Argument Retrieval for Comparative Questions.</figDesc><table><row><cell>Number</cell><cell>88</cell></row><row><cell>Title</cell><cell>Should I major in philosophy or psychology?</cell></row><row><cell>Objects</cell><cell>major in philosophy, psychology</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://webis.de/events.html?q=Touche#shared-tasks</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://www.args.me/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">http://commoncrawl.org</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">Three teams did not submit a paper describing their approach, though.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4">The expected format was also described at the lab's web page: https://webis.de/events/touche-22/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_5">https://webis.de/data.html#args-me-corpus</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_6">https://www.args.me/api-en.html</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_7">https://opennlp.apache.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_8">https://github.com/igorbrigadir/stopwords/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_9">https://connect.ebsco.com/s/article/What-are-stop-words-and-how-does-EBSCO-s-search-engine-handle-them?</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="11" xml:id="foot_10">https://www.semrush.com/blog/seo-stop-words/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="12" xml:id="foot_11">https://www.ranks.nl/stopwords</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="13" xml:id="foot_12">https://www.datamuse.com/api/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="14" xml:id="foot_13">https://99webtools.com/blog/list-of-english-stop-words/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="15" xml:id="foot_14">https://lemurproject.org/clueweb12/index.php</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="16" xml:id="foot_15">https://github.com/grill-lab/trec-cast-tools</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="17" xml:id="foot_16">Also available as a Python library: https://pypi.org/project/targer-api/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="18" xml:id="foot_17">http://www.dcs.gla.ac.uk/∼craigm/colbert.dnn.zip</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="19" xml:id="foot_18">https://github.com/MatthiasWinkelmann/english-words-names-brands-places</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="20" xml:id="foot_19">https://lemurproject.org/clueweb12/related-data.php</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="21" xml:id="foot_20">https://www.phash.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="22" xml:id="foot_21">Archived using https://github.com/webis-de/scriptor</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="23" xml:id="foot_22">Available at https://webis.de/data.html#touche22-image-retrieval-for-arguments</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="24" xml:id="foot_23">https://www.mturk.com</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="25" xml:id="foot_24">https://github.com/tesseract-ocr/tesseract</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="27" xml:id="foot_25">https://github.com/justinshenk/fer</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>We are very grateful to the CLEF 2022 organizers and the Touché participants, who allowed this lab to happen. We also want to thank our volunteer annotators who helped to create the relevance and argument quality assessments and our reviewers for their valuable feedback on the participants' notebooks.</p><p>This work was partially supported by the Deutsche Forschungsgemeinschaft (DFG) through the projects "ACQuA 2.0" (Answering Comparative Questions with Arguments; project number 376430233) and "OASiS" (Objective Argument Summarization in Search; project number 455913891) as part of the priority program "RATIO: Robust Argumentation Machines" (SPP 1999), and the German Ministry for Science and Education (BMBF) through the project "SharKI" (Shared Tasks as an Innovative Approach to Implement AI and Big Data-based Applications within Universities; grant FKZ 16DHB4021). We are also grateful to Jan Heinrich Reimer for developing the TARGER Python library and Erik Reuter for expanding a document collection for Task 2 with docT5query.</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A. Full Evaluation Results of Touché 2022: Argument Retrieval Table 6</head><p>Relevance results of all runs submitted to Task 1: Argument Retrieval for Controversial Questions. Reported are the mean nDCG@5 and the 95% confidence intervals. The baseline Swordsman is shown in bold. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 7</head><p>Quality results of all runs submitted to Task 1: Argument Retrieval for Controversial Questions. Reported are the mean nDCG@5 and the 95% confidence intervals. The baseline Swordsman is shown in bold. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 8</head><p>Coherence results of all runs submitted to Task 1: Argument Retrieval for Controversial Questions.</p><p>Reported are the mean nDCG@5 and the 95% confidence intervals. The baseline Swordsman is shown in bold. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 9</head><p>Relevance results of all runs submitted to Task 2: Argument Retrieval for Comparative Questions.</p><p>Reported are the mean nDCG@5 and the 95% confidence intervals; Puss in Boots baseline in bold. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Team</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 10</head><p>Quality results of all runs submitted to Task 2: Argument Retrieval for Comparative Questions. Reported are the mean nDCG@5 and the 95% confidence intervals; Puss in Boots baseline in bold. Reported are a macro-averaged F 1 for each team and run and number of documents N for which the stance was predicted; Puss in Boots baseline that always predicts 'no stance' is in bold. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Team</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 12</head><p>Results of all runs submitted to Task 3 Image Retrieval. Reported are the mean precision@10 (per stance) for topic relevance, argumentativeness, and stance relevance and the 95% confidence intervals (low and high). Results for the baseline are shown in bold. </p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Overview of Touché 2022: Argument retrieval</title>
		<author>
			<persName><forename type="first">A</forename><surname>Bondarenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kiesel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Syed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Gurcke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Beloucif</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Biemann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wachsmuth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction. 13th International Conference of the CLEF Association (CLEF 2022)</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<meeting><address><addrLine>Berlin Heidelberg New York</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Overview of Touché 2020: Argument retrieval</title>
		<author>
			<persName><forename type="first">A</forename><surname>Bondarenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Beloucif</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Gienapp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Ajjour</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Biemann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wachsmuth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<ptr target="http://ceur-ws.org/Vol-2696/" />
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2020 Evaluation Labs</title>
				<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">2696</biblScope>
		</imprint>
	</monogr>
	<note>CEUR Workshop Proceedings</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Overview of Touché 2021: Argument retrieval</title>
		<author>
			<persName><forename type="first">A</forename><surname>Bondarenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Gienapp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Beloucif</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Ajjour</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Biemann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wachsmuth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<ptr target="http://ceur-ws.org/Vol-2936/" />
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2021 Evaluation Labs</title>
				<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="volume">2936</biblScope>
		</imprint>
	</monogr>
	<note>CEUR Workshop Proceedings</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Okapi at TREC-3</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">E</forename><surname>Robertson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Walker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hancock-Beaulieu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gatford</surname></persName>
		</author>
		<ptr target="https://trec.nist.gov/pubs/trec3/papers/city.ps.gz" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of The Third Text REtrieval Conference, TREC 1994</title>
				<meeting>The Third Text REtrieval Conference, TREC 1994</meeting>
		<imprint>
			<publisher>NIST</publisher>
			<date type="published" when="1994">1994</date>
			<biblScope unit="page" from="109" to="126" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Exploring the limits of transfer learning with a unified text-to-text transformer</title>
		<author>
			<persName><forename type="first">C</forename><surname>Raffel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shazeer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Roberts</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Narang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Matena</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">J</forename><surname>Liu</surname></persName>
		</author>
		<ptr target="http://jmlr.org/papers/v21/20-074.html" />
	</analytic>
	<monogr>
		<title level="j">J. Mach. Learn. Res</title>
		<imprint>
			<biblScope unit="volume">21</biblScope>
			<biblScope unit="page">67</biblScope>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">Multitask prompted training enables zero-shot task generalization</title>
		<author>
			<persName><forename type="first">V</forename><surname>Sanh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Webson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Raffel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">H</forename><surname>Bach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Sutawika</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Alyafeai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Chaffin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Stiegler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">L</forename><surname>Scao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Raja</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Dey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">S</forename><surname>Bari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Thakker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sharma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Szczechla</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Chhablani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">V</forename><surname>Nayak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Datta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Jiang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Manica</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><forename type="middle">X</forename><surname>Yong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Pandey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Bawden</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Neeraj</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Rozen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Sharma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Santilli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Févry</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Fries</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Teehan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Biderman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Bers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Wolf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Rush</surname></persName>
		</author>
		<idno>CoRR abs/2110.08207</idno>
		<ptr target="https://arxiv.org/abs/2110.08207.arXiv:2110.08207" />
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">Pretrained Transformers for Text Ranking: BERT and Beyond, Synthesis Lectures on Human Language Technologies</title>
		<author>
			<persName><forename type="first">J</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Nogueira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Yates</surname></persName>
		</author>
		<idno type="DOI">10.2200/S01123ED1V01Y202108HLT053</idno>
		<ptr target="https://doi.org/10.2200/S01123ED1V01Y202108HLT053.doi:10.2200/S01123ED1V01Y202108HLT053" />
		<imprint>
			<date type="published" when="2021">2021</date>
			<publisher>Morgan &amp; Claypool Publishers</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">BERT: pre-training of deep bidirectional transformers for language understanding</title>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/n19-1423</idno>
		<ptr target="https://doi.org/10.18653/v1/n19-1423" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, ACL</title>
				<meeting>the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, ACL</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="4171" to="4186" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">Distilling dense representations for ranking using tightlycoupled teachers</title>
		<author>
			<persName><forename type="first">S</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lin</surname></persName>
		</author>
		<idno>CoRR abs/2010.11386</idno>
		<ptr target="https://arxiv.org/abs/2010.11386.arXiv:2010.11386" />
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Simple BM25 extension to multiple weighted fields</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">E</forename><surname>Robertson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zaragoza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Taylor</surname></persName>
		</author>
		<idno type="DOI">10.1145/1031171.1031181</idno>
		<ptr target="https://doi.org/10.1145/1031171.1031181" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 13th International Conference on Information and Knowledge Management, CIKM 2004</title>
				<meeting>the 13th International Conference on Information and Knowledge Management, CIKM 2004</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2004">2004</date>
			<biblScope unit="page" from="42" to="49" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">BEIR: A heterogeneous benchmark for zero-shot evaluation of information retrieval models</title>
		<author>
			<persName><forename type="first">N</forename><surname>Thakur</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Reimers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rücklé</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Srivastava</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
		<ptr target="https://openreview.net/forum?id=wCu6T5xFjeJ" />
	</analytic>
	<monogr>
		<title level="m">Thirty-fifth Conference on Neural Information Processing Systems Datasets and Benchmarks Track (Round 2)</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Simplified data wrangling with ir_datasets</title>
		<author>
			<persName><forename type="first">S</forename><surname>Macavaney</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Yates</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Feldman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Downey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Cohan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goharian</surname></persName>
		</author>
		<idno type="DOI">10.1145/3404835.3463254</idno>
		<ptr target="https://doi.org/10.1145/3404835.3463254" />
	</analytic>
	<monogr>
		<title level="m">Proceedings og the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2021</title>
				<meeting>og the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2021</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="2429" to="2436" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Retrieval of the best counterargument without prior topic knowledge</title>
		<author>
			<persName><forename type="first">H</forename><surname>Wachsmuth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Syed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<ptr target="https://www.aclweb.org/anthology/P18-1023/" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics, ACL 2018, Association for Computational Linguistics</title>
				<meeting>the 56th Annual Meeting of the Association for Computational Linguistics, ACL 2018, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="241" to="251" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Building an argument search engine for the web</title>
		<author>
			<persName><forename type="first">H</forename><surname>Wachsmuth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Al-Khatib</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Ajjour</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Puschmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Qu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Dorsch</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Morari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bevendorff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/w17-5106</idno>
		<ptr target="https://doi.org/10.18653/v1/w17-5106" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Fourth Workshop on Argument Mining (ArgMining), Association for Computational Linguistics</title>
				<meeting>the Fourth Workshop on Argument Mining (ArgMining), Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="49" to="59" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Identifying argumentative questions in web search logs</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Ajjour</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Braslavski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bondarenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<idno type="DOI">10.1145/3477495.3531864</idno>
	</analytic>
	<monogr>
		<title level="m">45th International ACM Conference on Research and Development in Information Retrieval (SIGIR 2022)</title>
				<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Data acquisition for argument search: The args.me corpus</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Ajjour</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wachsmuth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kiesel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-30179-8_4</idno>
		<ptr target="https://doi.org/10.1007/978-3-030-30179-8_4" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 42nd German Conference on AI, KI 2019</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<meeting>the 42nd German Conference on AI, KI 2019</meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">11793</biblScope>
			<biblScope unit="page" from="48" to="59" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Towards an argumentative content search engine using weak supervision</title>
		<author>
			<persName><forename type="first">R</forename><surname>Levy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Bogin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gretz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Aharonov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Slonim</surname></persName>
		</author>
		<ptr target="https://www.aclweb.org/anthology/C18-1176/" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 27th International Conference on Computational Linguistics, COLING 2018, Association for Computational Linguistics</title>
				<meeting>the 27th International Conference on Computational Linguistics, COLING 2018, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="2066" to="2081" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">ArgumenText: Searching for arguments in heterogeneous sources</title>
		<author>
			<persName><forename type="first">C</forename><surname>Stab</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Daxenberger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Stahlhut</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Miller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Schiller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Tauchmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Eger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
		<ptr target="https://www.aclweb.org/anthology/N18-5005" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics, NAACL 2018, Association for Computational Linguistics</title>
				<meeting>the 2018 Conference of the North American Chapter of the Association for Computational Linguistics, NAACL 2018, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="21" to="25" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m" type="main">On Rhetoric: A Theory of Civic Discourse</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">A</forename><surname>Aristotle</surname></persName>
		</author>
		<author>
			<persName><surname>Kennedy</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2006">2006</date>
			<publisher>Oxford University Press</publisher>
			<pubPlace>Oxford</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Argumentation quality assessment: Theory vs. practice</title>
		<author>
			<persName><forename type="first">H</forename><surname>Wachsmuth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Naderi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Habernal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Hou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Hirst</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/P17-2039</idno>
		<ptr target="https://doi.org/10.18653/v1/P17-2039" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics, ACL 2017, Association for Computational Linguistics</title>
				<meeting>the 55th Annual Meeting of the Association for Computational Linguistics, ACL 2017, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="250" to="255" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Argument search: Assessing argument relevance</title>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Gienapp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Euchner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Heilenkötter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Weidmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wachsmuth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<idno type="DOI">10.1145/3331184.3331327</idno>
		<ptr target="https://doi.org/10.1145/3331184.3331327" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 42nd International Conference on Research and Development in Information Retrieval, SIGIR 2019</title>
				<meeting>the 42nd International Conference on Research and Development in Information Retrieval, SIGIR 2019</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="1117" to="1120" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Efficient pairwise annotation of argument quality</title>
		<author>
			<persName><forename type="first">L</forename><surname>Gienapp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2020.acl-main.511</idno>
		<ptr target="https://aclanthology.org/2020.acl-main.511.doi:10.18653/v1/2020.acl-main.511" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics</title>
				<meeting>the 58th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="5772" to="5781" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">PageRank&quot; for argument relevance</title>
		<author>
			<persName><forename type="first">H</forename><surname>Wachsmuth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Ajjour</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/e17-1105</idno>
		<ptr target="https://doi.org/10.18653/v1/e17-1105" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics, EACL 2017, Association for Computational Linguistics</title>
				<meeting>the 15th Conference of the European Chapter of the Association for Computational Linguistics, EACL 2017, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="1117" to="1127" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<monogr>
		<title level="m" type="main">The PageRank Citation Ranking: Bringing Order to the Web</title>
		<author>
			<persName><forename type="first">L</forename><surname>Page</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Brin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Motwani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Winograd</surname></persName>
		</author>
		<idno>1999-66</idno>
		<ptr target="http://ilpubs.stanford.edu:8090/422/" />
		<imprint>
			<date type="published" when="1999">1999</date>
		</imprint>
		<respStmt>
			<orgName>Stanford InfoLab</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical Report</note>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">A framework for argument retrieval -ranking argument clusters by frequency and specificity</title>
		<author>
			<persName><forename type="first">L</forename><surname>Dumani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">J</forename><surname>Neumann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Schenkel</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-45439-5_29</idno>
		<ptr target="https://doi.org/10.1007/978-3-030-45439-5_29" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 42nd European Conference on IR Research (ECIR 2020)</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<meeting>the 42nd European Conference on IR Research (ECIR 2020)</meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">12035</biblScope>
			<biblScope unit="page" from="431" to="445" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Quality aware ranking of arguments</title>
		<author>
			<persName><forename type="first">L</forename><surname>Dumani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Schenkel</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-45439-5_29</idno>
		<ptr target="https://doi.org/10.1007/978-3-030-45439-5_29" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 29th ACM International Conference on Information &amp; Knowledge Management, CIKM &apos;20</title>
				<meeting>the 29th ACM International Conference on Information &amp; Knowledge Management, CIKM &apos;20</meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="335" to="344" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Computational argumentation quality assessment in natural language</title>
		<author>
			<persName><forename type="first">H</forename><surname>Wachsmuth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Naderi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Hou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Bilu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Prabhakaran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">A</forename><surname>Thijm</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Hirst</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<ptr target="http://aclweb.org/anthology/E17-1017" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics, EACL 2017</title>
				<meeting>the 15th Conference of the European Chapter of the Association for Computational Linguistics, EACL 2017</meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="176" to="187" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">A comparative web browser (CWB) for browsing and comparing web pages</title>
		<author>
			<persName><forename type="first">A</forename><surname>Nadamoto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Tanaka</surname></persName>
		</author>
		<idno type="DOI">10.1145/775152.775254</idno>
		<ptr target="https://doi.org/10.1145/775152.775254" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 12th International World Wide Web Conference, WWW 2003</title>
				<meeting>the 12th International World Wide Web Conference, WWW 2003</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2003">2003</date>
			<biblScope unit="page" from="727" to="735" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">CWS: A comparative web search system</title>
		<author>
			<persName><forename type="first">J</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zeng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Chen</surname></persName>
		</author>
		<idno type="DOI">10.1145/1135777.1135846</idno>
		<ptr target="https://doi.org/10.1145/1135777.1135846" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 15th International Conference on World Wide Web, WWW 2006</title>
				<meeting>the 15th International Conference on World Wide Web, WWW 2006</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="467" to="476" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">Identifying comparative sentences in text documents</title>
		<author>
			<persName><forename type="first">N</forename><surname>Jindal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Liu</surname></persName>
		</author>
		<idno type="DOI">10.1145/1148170.1148215</idno>
		<ptr target="https://doi.org/10.1145/1148170.1148215" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 29th Annual International Conference on Research and Development in Information Retrieval, SIGIR 2006</title>
				<meeting>the 29th Annual International Conference on Research and Development in Information Retrieval, SIGIR 2006</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="244" to="251" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Mining comparative sentences and relations</title>
		<author>
			<persName><forename type="first">N</forename><surname>Jindal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Liu</surname></persName>
		</author>
		<ptr target="http://www.aaai.org/Library/AAAI/2006/aaai06-209.php" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 21st National Conference on Artificial Intelligence and the 18th Innovative Applications of Artificial Intelligence Conference, AAAI 2006</title>
				<meeting>the 21st National Conference on Artificial Intelligence and the 18th Innovative Applications of Artificial Intelligence Conference, AAAI 2006</meeting>
		<imprint>
			<publisher>AAAI Press</publisher>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="1331" to="1336" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">A corpus of comparisons in product reviews</title>
		<author>
			<persName><forename type="first">W</forename><surname>Kessler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kuhn</surname></persName>
		</author>
		<ptr target="http://www.lrec-conf.org/proceedings/lrec2014/summaries/1001.html" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 9th International Conference on Language Resources and Evaluation, LREC 2014, European Language Resources Association (ELRA)</title>
				<meeting>the 9th International Conference on Language Resources and Evaluation, LREC 2014, European Language Resources Association (ELRA)</meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="2242" to="2248" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b32">
	<analytic>
		<title level="a" type="main">Categorizing comparative sentences</title>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bondarenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Franzek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Biemann</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/w19-4516</idno>
		<ptr target="https://doi.org/10.18653/v1/w19-4516" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 6th Workshop on Argument Mining, ArgMining@ACL 2019, Association for Computational Linguistics</title>
				<meeting>the 6th Workshop on Argument Mining, ArgMining@ACL 2019, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="136" to="145" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<analytic>
		<title level="a" type="main">Entity-aware dependency-based deep graph attention network for comparative preference classification</title>
		<author>
			<persName><forename type="first">N</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mazumder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Liu</surname></persName>
		</author>
		<ptr target="https://www.aclweb.org/anthology/2020.acl-main.512/" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Association for Computational Linguistics</title>
				<meeting>the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="5782" to="5788" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b34">
	<analytic>
		<title level="a" type="main">Answering comparative questions: Better than ten-blue-links?</title>
		<author>
			<persName><forename type="first">M</forename><surname>Schildwächter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bondarenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zenker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Biemann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<idno type="DOI">10.1145/3295750.3298916</idno>
		<ptr target="https://doi.org/10.1145/3295750.3298916" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference on Human Information Interaction and Retrieval</title>
				<meeting>the 2019 Conference on Human Information Interaction and Retrieval<address><addrLine>CHIIR</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2019">2019. 2019</date>
			<biblScope unit="page" from="361" to="365" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b35">
	<analytic>
		<title level="a" type="main">Which is better for deep learning: Python or matlab? answering comparative questions in natural language</title>
		<author>
			<persName><forename type="first">V</forename><surname>Chekalina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bondarenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Biemann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Beloucif</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Logacheva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<ptr target="https://www.aclweb.org/anthology/2021.eacl-demos.36/" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: System Demonstrations, EACL 2021, Association for Computational Linguistics</title>
				<meeting>the 16th Conference of the European Chapter of the Association for Computational Linguistics: System Demonstrations, EACL 2021, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="302" to="311" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b36">
	<analytic>
		<title level="a" type="main">Comparative web search questions</title>
		<author>
			<persName><forename type="first">A</forename><surname>Bondarenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Braslavski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Völske</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Aly</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Biemann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<idno type="DOI">10.1145/3336191.3371848</idno>
		<ptr target="https://dl.acm.org/doi/abs/10.1145/3336191.3371848" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 13th ACM International Conference on Web Search and Data Mining, WSDM 2020</title>
				<meeting>the 13th ACM International Conference on Web Search and Data Mining, WSDM 2020</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="52" to="60" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b37">
	<analytic>
		<title level="a" type="main">Towards understanding and answering comparative questions</title>
		<author>
			<persName><forename type="first">A</forename><surname>Bondarenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Ajjour</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Dittmar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Homann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Braslavski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<idno type="DOI">10.1145/3488560.3498534</idno>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 15th ACM International Conference on Web Search and Data Mining, WSDM 2022</title>
				<meeting>the 15th ACM International Conference on Web Search and Data Mining, WSDM 2022</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="66" to="74" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b38">
	<analytic>
		<title level="a" type="main">On images as evidence and arguments</title>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">J</forename><surname>Dove</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-94-007-4041-9_15</idno>
	</analytic>
	<monogr>
		<title level="m">Topical Themes in Argumentation Theory: Twenty Exploratory Studies, Argumentation Library</title>
				<editor>
			<persName><forename type="first">F</forename><forename type="middle">H</forename><surname>Van Eemeren</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Garssen</surname></persName>
		</editor>
		<meeting><address><addrLine>Netherlands; Dordrecht</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="223" to="238" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b39">
	<analytic>
		<title level="a" type="main">Images, emotions, politics</title>
		<author>
			<persName><forename type="first">F</forename><surname>Dunaway</surname></persName>
		</author>
		<idno type="DOI">10.1017/mah.2018.17</idno>
	</analytic>
	<monogr>
		<title level="j">Modern American History</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="369" to="376" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b40">
	<analytic>
		<title level="a" type="main">Visual argumentation: A further reappraisal</title>
		<author>
			<persName><forename type="first">G</forename><surname>Roque</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-94-007-4041-9_18</idno>
		<idno>doi:</idno>
		<ptr target="10.1007/978-94-007-4041-9_18" />
	</analytic>
	<monogr>
		<title level="m">Topical Themes in Argumentation Theory</title>
				<editor>
			<persName><forename type="first">F</forename><forename type="middle">H</forename><surname>Van Eemeren</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Garssen</surname></persName>
		</editor>
		<meeting><address><addrLine>Netherlands; Dordrecht</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2012">2012</date>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="page" from="273" to="288" />
		</imprint>
	</monogr>
	<note>series Title: Argumentation Library</note>
</biblStruct>

<biblStruct xml:id="b41">
	<analytic>
		<title level="a" type="main">Types of visual arguments, Argumentum</title>
		<author>
			<persName><forename type="first">I</forename><surname>Grancea</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of the Seminar of Discursive Logic, Argumentation Theory and Rhetoric</title>
		<imprint>
			<biblScope unit="volume">15</biblScope>
			<biblScope unit="page" from="16" to="34" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b42">
	<analytic>
		<title level="a" type="main">Martino, SemEval-2021 task 6: Detection of persuasion techniques in texts and images</title>
		<author>
			<persName><forename type="first">D</forename><surname>Dimitrov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Bin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ali</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Shaar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Alam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Silvestri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Firooz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Nakov</surname></persName>
		</author>
		<author>
			<persName><surname>Da San</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2021.semeval-1.7</idno>
		<ptr target="https://aclanthology.org/2021.semeval-1.7.doi:10.18653/v1/2021.semeval-1.7" />
	</analytic>
	<monogr>
		<title level="m">15th International Workshop on Semantic Evaluation (SemEval&apos;2021), Association for Computational Linguistics</title>
				<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="70" to="98" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b43">
	<analytic>
		<title level="a" type="main">Query-by-pictorial-example</title>
		<author>
			<persName><forename type="first">N</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Fu</surname></persName>
		</author>
		<idno type="DOI">10.1109/TSE.1980.230801</idno>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Software Engineering</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="page" from="519" to="524" />
			<date type="published" when="1980">1980</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b44">
	<analytic>
		<title level="a" type="main">Content-based representation and retrieval of visual media: A state-of-the-art review</title>
		<author>
			<persName><forename type="first">P</forename><surname>Aigrain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Petkovic</surname></persName>
		</author>
		<idno type="DOI">10.1007/BF00393937</idno>
	</analytic>
	<monogr>
		<title level="j">Multimedia Tools and Applications</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="179" to="202" />
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b45">
	<analytic>
		<title level="a" type="main">Content-based image retrieval and feature extraction: A comprehensive review</title>
		<author>
			<persName><forename type="first">A</forename><surname>Latif</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rasheed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Sajid</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ahmed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ali</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">I</forename><surname>Ratyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Zafar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">H</forename><surname>Dar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sajid</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Khalil</surname></persName>
		</author>
		<idno type="DOI">10.1155/2019/9658350</idno>
	</analytic>
	<monogr>
		<title level="j">Mathematical Problems in Engineering</title>
		<imprint>
			<biblScope unit="page">21</biblScope>
			<date type="published" when="2019">2019. 2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b46">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Wu</surname></persName>
		</author>
		<ptr target="https://support.google.com/webmasters/answer/114016" />
		<title level="m">Learn more about what you see on google images, Google Blog</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b47">
	<monogr>
		<ptr target="https://support.google.com/webmasters/answer/114016" />
		<title level="m">Google, Google images best practices, Google Developers</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b48">
	<analytic>
		<title level="a" type="main">A survey on emotional semantic image retrieval</title>
		<author>
			<persName><forename type="first">W</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>He</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICIP.2008.4711705</idno>
	</analytic>
	<monogr>
		<title level="m">International Conference on Image Processing (ICIP 2008</title>
				<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2008">2008</date>
			<biblScope unit="page" from="117" to="120" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b49">
	<analytic>
		<title level="a" type="main">Color emotions for multi-colored images</title>
		<author>
			<persName><forename type="first">M</forename><surname>Solli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Lenz</surname></persName>
		</author>
		<idno type="DOI">10.1002/col.20604</idno>
	</analytic>
	<monogr>
		<title level="j">Color Research &amp; Application</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<biblScope unit="page" from="210" to="221" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b50">
	<analytic>
		<title level="a" type="main">Image retrieval for arguments using stance-aware query expansion</title>
		<author>
			<persName><forename type="first">J</forename><surname>Kiesel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Reichenbach</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 8th Workshop on Argument Mining, ArgMining 2021 at EMNLP, ACL</title>
				<meeting>the 8th Workshop on Argument Mining, ArgMining 2021 at EMNLP, ACL</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="36" to="45" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b51">
	<analytic>
		<title level="a" type="main">TIRA integrated research architecture</title>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Gollub</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wiegmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-22948-1_5</idno>
		<ptr target="https://doi.org/10.1007/978-3-030-22948-1_5" />
	</analytic>
	<monogr>
		<title level="m">Information Retrieval Evaluation in a Changing World -Lessons Learned from 20 Years of CLEF</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">41</biblScope>
			<biblScope unit="page" from="123" to="160" />
		</imprint>
	</monogr>
	<note>The Information Retrieval Series</note>
</biblStruct>

<biblStruct xml:id="b52">
	<analytic>
		<title level="a" type="main">Extractive snippet generation for arguments</title>
		<author>
			<persName><forename type="first">M</forename><surname>Alshomary</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Düsterhus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wachsmuth</surname></persName>
		</author>
		<idno type="DOI">10.1145/3397271.3401186</idno>
		<ptr target="https://doi.org/10.1145/3397271.3401186" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 43nd International ACM Conference on Research and Development in Information Retrieval, SIGIR 2020</title>
				<meeting>the 43nd International ACM Conference on Research and Development in Information Retrieval, SIGIR 2020</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="1969" to="1972" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b53">
	<analytic>
		<title level="a" type="main">TrecTools: an Open-source Python Library for Information Retrieval Practitioners Involved in TREC-like Campaigns</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R M</forename><surname>Palotti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Scells</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Zuccon</surname></persName>
		</author>
		<idno type="DOI">10.1145/3331184.3331399</idno>
		<ptr target="https://doi.org/10.1145/3331184.3331399" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 42nd International Conference on Research and Development in Information Retrieval, SIGIR 2019</title>
				<meeting>the 42nd International Conference on Research and Development in Information Retrieval, SIGIR 2019</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="1325" to="1328" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b54">
	<analytic>
		<title level="a" type="main">Project debater apis: Decomposing the AI grand challenge</title>
		<author>
			<persName><forename type="first">R</forename><surname>Bar-Haim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Kantor</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Venezian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Katz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Slonim</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2021.emnlp-demo.31</idno>
		<ptr target="https://doi.org/10.18653/v1/2021.emnlp-demo.31" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, EMNLP 2021, Online and</title>
				<meeting>the 2021 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, EMNLP 2021, Online and<address><addrLine>Punta Cana, Dominican Republic</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021-11-11">7-11 November, 2021. 2021</date>
			<biblScope unit="page" from="267" to="274" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b55">
	<analytic>
		<title level="a" type="main">Classification and clustering of arguments with contextualized word embeddings</title>
		<author>
			<persName><forename type="first">N</forename><surname>Reimers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Schiller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Beck</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Daxenberger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Stab</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/P19-1054</idno>
		<ptr target="https://aclanthology.org/P19-1054.doi:10.18653/v1/P19-1054" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics</title>
				<meeting>the 57th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics<address><addrLine>Florence, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="567" to="578" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b56">
	<analytic>
		<title level="a" type="main">A large-scale dataset for argument quality ranking: Construction and analysis</title>
		<author>
			<persName><forename type="first">S</forename><surname>Gretz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Friedman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Cohen-Karlik</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Toledo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Lahav</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Aharonov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Slonim</surname></persName>
		</author>
		<ptr target="https://ojs.aaai.org/index.php/AAAI/article/view/6285" />
	</analytic>
	<monogr>
		<title level="m">The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020</title>
				<meeting><address><addrLine>EAAI</addrLine></address></meeting>
		<imprint>
			<publisher>AAAI Press</publisher>
			<date type="published" when="2020">2020. 2020</date>
			<biblScope unit="page" from="7805" to="7813" />
		</imprint>
	</monogr>
	<note>The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence</note>
</biblStruct>

<biblStruct xml:id="b57">
	<analytic>
		<title level="a" type="main">Distributed representations of sentences and documents</title>
		<author>
			<persName><forename type="first">Q</forename><surname>Le</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Mikolov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International conference on machine learning</title>
				<meeting><address><addrLine>PMLR</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="1188" to="1196" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b58">
	<analytic>
		<title level="a" type="main">spacy 2: Natural language understanding with bloom embeddings, convolutional neural networks and incremental parsing</title>
		<author>
			<persName><forename type="first">M</forename><surname>Honnibal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Montani</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">To appear</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="page" from="411" to="420" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b59">
	<analytic>
		<title level="a" type="main">Sentence-BERT: Sentence embeddings using siamese bertnetworks</title>
		<author>
			<persName><forename type="first">N</forename><surname>Reimers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/D19-1410</idno>
		<ptr target="https://doi.org/10.18653/v1/D19-1410" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP 2019, Association for Computational Linguistics</title>
				<meeting>the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing, EMNLP-IJCNLP 2019, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="3980" to="3990" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b60">
	<analytic>
		<title level="a" type="main">Language models are unsupervised multitask learners</title>
		<author>
			<persName><forename type="first">A</forename><surname>Radford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Child</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Luan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Amodei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sutskever</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">OpenAI blog</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page">9</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b61">
	<analytic>
		<title level="a" type="main">Using BERT to retrieve relevant and argumentative sentence pairs</title>
		<author>
			<persName><forename type="first">P</forename><surname>Sülzle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Wenzlitschke</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2022 Evaluation Labs, CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b62">
	<analytic>
		<title level="a" type="main">SE-UPD@CLEF: Team INTSEG on argument retrieval for controversial questions</title>
		<author>
			<persName><forename type="first">S</forename><surname>Bahrami</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">P</forename><surname>Goli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Pasin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Rajkumari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">M</forename><surname>Sohail</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Tahan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2022 Evaluation Labs, CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b63">
	<analytic>
		<title level="a" type="main">Team Bruce Banner at Touché 2022: Argument retrieval for controversial questions</title>
		<author>
			<persName><forename type="first">B</forename><surname>Moreira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Cardoso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Martins</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Goularte</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2022 Evaluation Labs, CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b64">
	<analytic>
		<title level="a" type="main">Pyserini: A python toolkit for reproducible information retrieval research with sparse and dense representations</title>
		<author>
			<persName><forename type="first">J</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Pradeep</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Nogueira</surname></persName>
		</author>
		<idno type="DOI">10.1145/3404835.3463238</idno>
		<ptr target="https://doi.org/10.1145/3404835.3463238" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2021</title>
				<meeting>the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2021</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="2356" to="2362" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b65">
	<analytic>
		<title level="a" type="main">SEUPD@CLEF: Team 6musk on argument retrieval for controversial questions by using pairs selection and query expansion</title>
		<author>
			<persName><forename type="first">L</forename><surname>Cappellotto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lando</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Lupu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Mariotto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Rosalen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2022 Evaluation Labs, CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b66">
	<analytic>
		<title level="a" type="main">An algorithm for suffix stripping</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">F</forename><surname>Porter</surname></persName>
		</author>
		<idno type="DOI">10.1108/eb046814</idno>
		<ptr target="https://doi.org/10.1108/eb046814" />
	</analytic>
	<monogr>
		<title level="j">Program</title>
		<imprint>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="page" from="130" to="137" />
			<date type="published" when="1980">1980</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b67">
	<analytic>
		<title level="a" type="main">Viewing morphology as an inference process</title>
		<author>
			<persName><forename type="first">R</forename><surname>Krovetz</surname></persName>
		</author>
		<idno type="DOI">10.1145/160688.160718</idno>
		<idno>doi:10.1145/160688.160718</idno>
		<ptr target="https://doi.org/10.1145/160688.160718" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 16th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR &apos;93</title>
				<meeting>the 16th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR &apos;93<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="1993">1993</date>
			<biblScope unit="page" from="191" to="202" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b68">
	<analytic>
		<title level="a" type="main">RCV1: A new benchmark collection for text categorization research</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">D</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">G</forename><surname>Rose</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Li</surname></persName>
		</author>
		<ptr target="http://jmlr.org/papers/volume5/lewis04a/lewis04a.pdf" />
	</analytic>
	<monogr>
		<title level="j">J. Mach. Learn. Res</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="page" from="361" to="397" />
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b69">
	<analytic>
		<title level="a" type="main">WordNet: A lexical database for English</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">A</forename><surname>Miller</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Communications of the ACM</title>
		<imprint>
			<biblScope unit="volume">38</biblScope>
			<biblScope unit="page" from="39" to="41" />
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b70">
	<analytic>
		<title level="a" type="main">Efficient estimation of word representations in vector space</title>
		<author>
			<persName><forename type="first">T</forename><surname>Mikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Corrado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Dean</surname></persName>
		</author>
		<ptr target="http://arxiv.org/abs/1301.3781" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 1st International Conference on Learning Representations, ICLR 2013</title>
				<meeting>the 1st International Conference on Learning Representations, ICLR 2013</meeting>
		<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b71">
	<analytic>
		<title level="a" type="main">SE-UPD@CLEF: Team Gamora on argument retrieval for controversial questions</title>
		<author>
			<persName><forename type="first">A</forename><surname>Benetti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">D</forename><surname>Togni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Foti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Lacini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Matteazzi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Sgarbossa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2022 Evaluation Labs, CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b72">
	<analytic>
		<title level="a" type="main">Glove: Global vectors for word representation</title>
		<author>
			<persName><forename type="first">J</forename><surname>Pennington</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Socher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
		<idno type="DOI">10.3115/v1/d14-1162</idno>
		<ptr target="https://doi.org/10.3115/v1/d14-1162" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, EMNLP 2014</title>
				<editor>
			<persName><forename type="first">A</forename><surname>Moschitti</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Pang</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">W</forename><surname>Daelemans</surname></persName>
		</editor>
		<meeting>the 2014 Conference on Empirical Methods in Natural Language Processing, EMNLP 2014<address><addrLine>Doha, Qatar</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2014">October 25-29, 2014. 2014</date>
			<biblScope unit="page" from="1532" to="1543" />
		</imprint>
	</monogr>
	<note>, A meeting of SIGDAT, a Special Interest Group of the ACL, ACL</note>
</biblStruct>

<biblStruct xml:id="b73">
	<analytic>
		<title level="a" type="main">SEUPD@CLEF: Team Lgtm on argument retrieval for controversial questions</title>
		<author>
			<persName><forename type="first">M</forename><surname>Barusco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">D</forename><surname>Fiume</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Forzan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">G</forename><surname>Peloso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Rizzetto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Soleymani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2022 Evaluation Labs, CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b74">
	<analytic>
		<title level="a" type="main">How effective is suffixing?</title>
		<author>
			<persName><forename type="first">D</forename><surname>Harman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of the american society for information science</title>
		<imprint>
			<biblScope unit="volume">42</biblScope>
			<biblScope unit="page" from="7" to="15" />
			<date type="published" when="1991">1991</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b75">
	<analytic>
		<title level="a" type="main">The stanford corenlp natural language processing toolkit</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Surdeanu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bauer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R</forename><surname>Finkel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Bethard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Mcclosky</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of 52nd annual meeting of the association for computational linguistics: system demonstrations</title>
				<meeting>52nd annual meeting of the association for computational linguistics: system demonstrations</meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="55" to="60" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b76">
	<analytic>
		<title level="a" type="main">Automatic keyword extraction from individual documents</title>
		<author>
			<persName><forename type="first">S</forename><surname>Rose</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Engel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Cramer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Cowley</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Text mining: applications and theory</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="10" to="1002" />
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b77">
	<analytic>
		<title level="a" type="main">Vader: A parsimonious rule-based model for sentiment analysis of social media text</title>
		<author>
			<persName><forename type="first">C</forename><surname>Hutto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Gilbert</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the international AAAI conference on web and social media</title>
				<meeting>the international AAAI conference on web and social media</meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="216" to="225" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b78">
	<monogr>
		<title level="m" type="main">Derivation of new readability formulas (automated readability index, fog count and flesch reading ease formula) for navy enlisted personnel</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">P</forename><surname>Kincaid</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">P</forename><surname>Fishburne</surname><genName>Jr</genName></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">L</forename><surname>Rogers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">S</forename><surname>Chissom</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1975">1975</date>
		</imprint>
		<respStmt>
			<orgName>Naval Technical Training Command Millington TN Research Branch</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical Report</note>
</biblStruct>

<biblStruct xml:id="b79">
	<analytic>
		<title level="a" type="main">SEUPD@CLEF: SPAM on argument retrieval for controversial questions</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">S</forename><surname>Ebrahimi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Crivellari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Sah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hansen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mehrbanou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Ashok</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2022 Evaluation Labs, CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b80">
	<analytic>
		<title level="a" type="main">Similar but different: Simple re-ranking approaches for argument retrieval</title>
		<author>
			<persName><forename type="first">J</forename><surname>Wuerf</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2022 Evaluation Labs, CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b81">
	<analytic>
		<title level="a" type="main">The use of mmr, diversity-based reranking for reordering documents and producing summaries</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">G</forename><surname>Carbonell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Goldstein</surname></persName>
		</author>
		<idno type="DOI">10.1145/290941.291025</idno>
		<ptr target="https://doi.org/10.1145/290941.291025.doi:10.1145/290941.291025" />
	</analytic>
	<monogr>
		<title level="m">SIGIR &apos;98: Proceedings of the 21st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval</title>
				<editor>
			<persName><forename type="first">W</forename><forename type="middle">B</forename><surname>Croft</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Moffat</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><forename type="middle">J</forename><surname>Van Rijsbergen</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Wilkinson</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Zobel</surname></persName>
		</editor>
		<meeting><address><addrLine>Melbourne, Australia</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="1998">August 24-28 1998. 1998</date>
			<biblScope unit="page" from="335" to="336" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b82">
	<analytic>
		<title level="a" type="main">From word embeddings to document distances</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Kusner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">I</forename><surname>Kolkin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">Q</forename><surname>Weinberger</surname></persName>
		</author>
		<ptr target="http://proceedings.mlr.press/v37/kusnerb15.html" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 32nd International Conference on Machine Learning, ICML 2015</title>
				<editor>
			<persName><forename type="first">F</forename><forename type="middle">R</forename><surname>Bach</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">D</forename><forename type="middle">M</forename><surname>Blei</surname></persName>
		</editor>
		<meeting>the 32nd International Conference on Machine Learning, ICML 2015<address><addrLine>Lille, France</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2015-07-11">6-11 July 2015. 2015</date>
			<biblScope unit="volume">37</biblScope>
			<biblScope unit="page" from="957" to="966" />
		</imprint>
	</monogr>
	<note>JMLR Workshop and Conference Proceedings</note>
</biblStruct>

<biblStruct xml:id="b83">
	<analytic>
		<title level="a" type="main">Finding pairs of argumentative sentences using embeddings</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">V</forename><surname>Ta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Reiner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Von Detten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Stöhr</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2022 Evaluation Labs, CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b84">
	<analytic>
		<title level="a" type="main">Two-stage retrieval for pairs of argumentative sentences</title>
		<author>
			<persName><forename type="first">S</forename><surname>Schmidt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Probst</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Bartelt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hinz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2022 Evaluation Labs, CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b85">
	<analytic>
		<title level="a" type="main">A study of smoothing methods for language models applied to ad hoc information retrieval</title>
		<author>
			<persName><forename type="first">C</forename><surname>Zhai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">D</forename><surname>Lafferty</surname></persName>
		</author>
		<idno type="DOI">10.1145/383952.384019</idno>
		<ptr target="https://doi.org/10.1145/383952.384019" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 24th International Conference on Research and Development in Information Retrieval, SIGIR 2001</title>
				<meeting>the 24th International Conference on Research and Development in Information Retrieval, SIGIR 2001</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2001">2001</date>
			<biblScope unit="page" from="334" to="342" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b86">
	<analytic>
		<title level="a" type="main">Frequentist and bayesian approach to information retrieval</title>
		<author>
			<persName><forename type="first">G</forename><surname>Amati</surname></persName>
		</author>
		<idno type="DOI">10.1007/11735106_3</idno>
		<ptr target="https://doi.org/10.1007/11735106_3.doi:10.1007/11735106\_3" />
	</analytic>
	<monogr>
		<title level="m">Advances in Information Retrieval, 28th European Conference on IR Research, ECIR 2006</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">M</forename><surname>Lalmas</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Macfarlane</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Rüger</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Tombros</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Tsikrika</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Yavlinsky</surname></persName>
		</editor>
		<meeting><address><addrLine>London, UK</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2006">April 10-12, 2006. 2006</date>
			<biblScope unit="volume">3936</biblScope>
			<biblScope unit="page" from="13" to="24" />
		</imprint>
	</monogr>
	<note>Proceedings</note>
</biblStruct>

<biblStruct xml:id="b87">
	<analytic>
		<title level="a" type="main">Declarative experimentation in information retrieval using pyterrier</title>
		<author>
			<persName><forename type="first">C</forename><surname>Macdonald</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Tonellotto</surname></persName>
		</author>
		<idno type="DOI">10.1145/3409256.3409829</idno>
		<idno>doi:10.1145/3409256.3409829</idno>
		<ptr target="https://doi.org/10.1145/3409256.3409829" />
	</analytic>
	<monogr>
		<title level="m">ICTIR &apos;20: The 2020 ACM SIGIR International Conference on the Theory of Information Retrieval, Virtual Event</title>
				<editor>
			<persName><forename type="first">K</forename><surname>Balog</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">V</forename><surname>Setty</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Lioma</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Zhang</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">K</forename><surname>Berberich</surname></persName>
		</editor>
		<meeting><address><addrLine>, Norway</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2020">September 14-17, 2020. 2020</date>
			<biblScope unit="page" from="161" to="168" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b88">
	<analytic>
		<title level="a" type="main">Efficient pairwise annotation of argument quality</title>
		<author>
			<persName><forename type="first">L</forename><surname>Gienapp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<ptr target="https://www.aclweb.org/anthology/2020.acl-main.511/" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Association for Computational Linguistics</title>
				<meeting>the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="5772" to="5781" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b89">
	<analytic>
		<title level="a" type="main">Probabilistic models of information retrieval based on measuring the divergence from randomness</title>
		<author>
			<persName><forename type="first">G</forename><surname>Amati</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">J</forename><surname>Van Rijsbergen</surname></persName>
		</author>
		<idno type="DOI">10.1145/582415.582416</idno>
		<idno>doi:10.1145/582415.582416</idno>
		<ptr target="http://doi.acm.org/10.1145/582415.582416" />
	</analytic>
	<monogr>
		<title level="j">ACM Trans. Inf. Syst</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="page" from="357" to="389" />
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b90">
	<analytic>
		<title level="a" type="main">Passage retrieval revisited</title>
		<author>
			<persName><forename type="first">M</forename><surname>Kaszkiel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zobel</surname></persName>
		</author>
		<idno type="DOI">10.1145/258525.258561</idno>
		<ptr target="https://doi.org/10.1145/258525.258561" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 20th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 1997</title>
				<editor>
			<persName><forename type="first">N</forename><forename type="middle">J</forename><surname>Belkin</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">D</forename><surname>Narasimhalu</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Willett</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">W</forename><forename type="middle">R</forename><surname>Hersh</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Can</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><forename type="middle">M</forename><surname>Voorhees</surname></persName>
		</editor>
		<meeting>the 20th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 1997<address><addrLine>Philadelphia, PA, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="1997">July 27-31, 1997. 1997</date>
			<biblScope unit="page" from="178" to="185" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b91">
	<analytic>
		<title level="a" type="main">Copycat: Near-duplicates within and between the clueweb and the common crawl</title>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bevendorff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Gienapp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Völske</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<idno type="DOI">10.1145/3404835.3463246</idno>
		<ptr target="https://dl.acm.org/doi/10.1145/3404835.3463246" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 44th International ACM Conference on Research and Development in Information Retrieval, SIGIR 2021</title>
				<meeting>the 44th International ACM Conference on Research and Development in Information Retrieval, SIGIR 2021</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="2398" to="2404" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b92">
	<analytic>
		<title level="a" type="main">Sampling bias due to nearduplicates in learning to rank</title>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bevendorff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Reimer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<idno type="DOI">10.1145/3397271.3401212</idno>
		<ptr target="https://dl.acm.org/doi/10.1145/3397271.3401212" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 43rd International ACM Conference on Research and Development in Information Retrieval, SIGIR 2020</title>
				<meeting>the 43rd International ACM Conference on Research and Development in Information Retrieval, SIGIR 2020</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="1997" to="2000" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b93">
	<analytic>
		<title level="a" type="main">The effect of content-equivalent nearduplicates on the evaluation of search engines</title>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bittner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<ptr target="https://link.springer.com/chapter/10.1007%2F978-3-030-45442-5_2" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 42nd European Conference on IR Research (ECIR 2020)</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<meeting>the 42nd European Conference on IR Research (ECIR 2020)</meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">12036</biblScope>
			<biblScope unit="page" from="12" to="19" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b94">
	<monogr>
		<author>
			<persName><forename type="first">R</forename><surname>Nogueira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Epistemic</surname></persName>
		</author>
		<ptr target="https://cs.uwaterloo.ca/~jimmylin/publications/Nogueira_Lin_2019_docTTTTTquery-v2.pdf" />
		<title level="m">From doc2query to doctttttquery</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note>Online preprint</note>
</biblStruct>

<biblStruct xml:id="b95">
	<analytic>
		<title level="a" type="main">MS MARCO: A human generated MAchine Reading COmprehension dataset</title>
		<author>
			<persName><forename type="first">T</forename><surname>Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Rosenberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Tiwary</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Majumder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Deng</surname></persName>
		</author>
		<ptr target="http://ceur-ws.org/Vol-1773/CoCoNIPS_2016_paper9.pdf" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Workshop on Cognitive Computation: Integrating Neural and Symbolic Approaches 2016 at NIPS</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<meeting>the Workshop on Cognitive Computation: Integrating Neural and Symbolic Approaches 2016 at NIPS</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="volume">1773</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b96">
	<analytic>
		<title level="a" type="main">ALBERT: A lite BERT for self-supervised learning of language representations</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Lan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Goodman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Gimpel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Sharma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Soricut</surname></persName>
		</author>
		<ptr target="https://openreview.net/forum?id=H1eA7AEtvS" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 8th International Conference on Learning Representations, ICLR 2020</title>
				<meeting>the 8th International Conference on Learning Representations, ICLR 2020<address><addrLine>OpenReview</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b97">
	<analytic>
		<title level="a" type="main">TARGER: Neural argument mining at your fingertips</title>
		<author>
			<persName><forename type="first">A</forename><surname>Chernodub</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Oliynyk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Heidenreich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bondarenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Biemann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/p19-3031</idno>
		<ptr target="https://doi.org/10.18653/v1/p19-3031" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, ACL 2019, ACL</title>
				<meeting>the 57th Annual Meeting of the Association for Computational Linguistics, ACL 2019, ACL</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="195" to="200" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b98">
	<analytic>
		<title level="a" type="main">ColBERT: efficient and effective passage search via contextualized late interaction over BERT</title>
		<author>
			<persName><forename type="first">O</forename><surname>Khattab</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Zaharia</surname></persName>
		</author>
		<idno type="DOI">10.1145/3397271.3401075</idno>
		<ptr target="https://doi.org/10.1145/3397271.3401075" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 43rd International ACM SIGIR conference on research and development in Information Retrieval, SIGIR 2020</title>
				<editor>
			<persName><forename type="first">J</forename><surname>Huang</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>Chang</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">X</forename><surname>Cheng</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Kamps</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">V</forename><surname>Murdock</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Wen</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</editor>
		<meeting>the 43rd International ACM SIGIR conference on research and development in Information Retrieval, SIGIR 2020</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="39" to="48" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b99">
	<monogr>
		<title level="m" type="main">The expando-mono-duo design pattern for text ranking with pretrained sequence-to-sequence models</title>
		<author>
			<persName><forename type="first">R</forename><surname>Pradeep</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Nogueira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lin</surname></persName>
		</author>
		<idno>CoRR abs/2101.05667</idno>
		<ptr target="https://arxiv.org/abs/2101.05667.arXiv:2101.05667" />
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b100">
	<analytic>
		<title level="a" type="main">LeviRank: Limited query expansion with voting integration for document retrieval and ranking</title>
		<author>
			<persName><forename type="first">A</forename><surname>Rana</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Golchha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Juntunen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Coajă</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Elzamarany</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-C</forename><surname>Hung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">P</forename><surname>Ponzetto</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2022 Evaluation Labs, CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b101">
	<monogr>
		<title level="m" type="main">Natural Language Processing with Python</title>
		<author>
			<persName><forename type="first">S</forename><surname>Bird</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Klein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Loper</surname></persName>
		</author>
		<ptr target="http://www.oreilly.de/catalog/9780596516499/index.html" />
		<imprint>
			<date type="published" when="2009">2009</date>
			<publisher>O&apos;Reilly</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b102">
	<monogr>
		<title level="m" type="main">RoBERTa: A robustly optimized BERT pretraining approach</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ott</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Joshi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Levy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Stoyanov</surname></persName>
		</author>
		<idno>CoRR abs/1907.11692</idno>
		<ptr target="http://arxiv.org/abs/1907.11692.arXiv:1907.11692" />
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b103">
	<analytic>
		<title level="a" type="main">A broad-coverage challenge corpus for sentence understanding through inference</title>
		<author>
			<persName><forename type="first">A</forename><surname>Williams</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Nangia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">R</forename><surname>Bowman</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/n18-1101</idno>
		<ptr target="https://doi.org/10.18653/v1/n18-1101" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2018, Association for Computational Linguistics</title>
				<meeting>the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2018, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="1112" to="1122" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b104">
	<analytic>
		<title level="a" type="main">Argument retrieval for comparative question</title>
		<author>
			<persName><forename type="first">M</forename><surname>Aba</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Azra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gallo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Mohammad</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Piacere</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Virginio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Aldo</forename><surname>Nadi</surname></persName>
		</author>
		<author>
			<persName><surname>Touché</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2022 Evaluation Labs, CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022. 2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b105">
	<monogr>
		<title level="m" type="main">Snowball: A language for stemming algorithms</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">F</forename><surname>Porter</surname></persName>
		</author>
		<ptr target="http://snowball.tartarus.org/texts/introduction.html" />
		<imprint>
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b106">
	<monogr>
		<title level="m" type="main">Relevance feedback in information retrieval, The Smart retrieval systemexperiments in automatic document processing</title>
		<author>
			<persName><forename type="first">J</forename><surname>Rocchio</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1971">1971</date>
			<biblScope unit="page" from="313" to="323" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b107">
	<analytic>
		<title level="a" type="main">Reciprocal rank fusion outperforms condorcet and individual rank learning methods</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">V</forename><surname>Cormack</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">L A</forename><surname>Clarke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Büttcher</surname></persName>
		</author>
		<idno type="DOI">10.1145/1571941.1572114</idno>
		<ptr target="https://doi.org/10.1145/1571941.1572114" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 32nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2009</title>
				<meeting>the 32nd Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR 2009</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="758" to="759" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b108">
	<analytic>
		<title level="a" type="main">Retrieving comparative arguments using deep language models</title>
		<author>
			<persName><forename type="first">V</forename><surname>Chekalina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2022 Evaluation Labs, CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b109">
	<analytic>
		<title level="a" type="main">SEUPD@CLEF: Team Hextech on argument retrieval for comparative questions. the importance of adjectives in documents quality evaluation</title>
		<author>
			<persName><forename type="first">A</forename><surname>Chimetto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Peressoni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Sabbatini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Tommasin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Varotto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Zanardelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2022 Evaluation Labs, CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b110">
	<analytic>
		<title level="a" type="main">Development of a stemming algorithm</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">B</forename><surname>Lovins</surname></persName>
		</author>
		<ptr target="http://www.mt-archive.info/MT-1968-Lovins.pdf" />
	</analytic>
	<monogr>
		<title level="j">Mech. Transl. Comput. Linguistics</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="page" from="22" to="31" />
			<date type="published" when="1968">1968</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b111">
	<analytic>
		<title level="a" type="main">Stacked model based argument extraction and stance detection using embedded LSTM model</title>
		<author>
			<persName><forename type="first">P</forename><surname>Rajula</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-C</forename><surname>Hung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">P</forename><surname>Ponzetto</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2022 Evaluation Labs, CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b112">
	<monogr>
		<title level="m" type="main">Distilbert, a distilled version of BERT: smaller, faster, cheaper and lighter</title>
		<author>
			<persName><forename type="first">V</forename><surname>Sanh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Debut</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chaumond</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Wolf</surname></persName>
		</author>
		<idno>CoRR abs/1910.01108</idno>
		<ptr target="http://arxiv.org/abs/1910.01108.arXiv:1910.01108" />
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b113">
	<analytic>
		<title level="a" type="main">Distilbert-based argumentation retrieval for answering comparative questions</title>
		<author>
			<persName><forename type="first">A</forename><surname>Alhamzeh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bouhaouel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Egyed-Zsigmond</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Mitrovic</surname></persName>
		</author>
		<idno>WS.org</idno>
		<ptr target="http://ceur-ws.org/Vol-2936/paper-209.pdf" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Working Notes of CLEF 2021 -Conference and Labs of the Evaluation Forum</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<meeting>the Working Notes of CLEF 2021 -Conference and Labs of the Evaluation Forum</meeting>
		<imprint>
			<date type="published" when="2021">2936. 2021</date>
			<biblScope unit="page" from="2319" to="2330" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b114">
	<analytic>
		<title level="a" type="main">Grimjack at Touché 2022: Axiomatic re-ranking and query reformulation</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">H</forename><surname>Reimer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Huck</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bondarenko</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2022 Evaluation Labs, CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b115">
	<analytic>
		<title level="a" type="main">Webis at TREC 2019: Decision track</title>
		<author>
			<persName><forename type="first">A</forename><surname>Bondarenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Kasturia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Völske</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 28th International Text Retrieval Conference, TREC 2019</title>
				<editor>
			<persName><forename type="first">E</forename><surname>Voorhees</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Ellis</surname></persName>
		</editor>
		<meeting>the 28th International Text Retrieval Conference, TREC 2019</meeting>
		<imprint>
			<publisher>NIST</publisher>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b116">
	<analytic>
		<title level="a" type="main">Webis at TREC 2020: Health Misinformation track</title>
		<author>
			<persName><forename type="first">J</forename><surname>Bevendorff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bondarenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Günther</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Völske</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 29th International Text Retrieval Conference, TREC 2020</title>
				<editor>
			<persName><forename type="first">E</forename><surname>Voorhees</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Ellis</surname></persName>
		</editor>
		<meeting>the 29th International Text Retrieval Conference, TREC 2020</meeting>
		<imprint>
			<publisher>NIST</publisher>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b117">
	<analytic>
		<title level="a" type="main">Quality-aware argument re-ranking for comparative questions</title>
		<author>
			<persName><forename type="first">P</forename><surname>Rösner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Arnhold</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Xylander</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2022 Evaluation Labs, CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b118">
	<analytic>
		<title level="a" type="main">LexRank: Graph-based lexical centrality as salience in text summarization</title>
		<author>
			<persName><forename type="first">G</forename><surname>Erkan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">R</forename><surname>Radev</surname></persName>
		</author>
		<idno type="DOI">10.1613/jair.1523</idno>
		<ptr target="https://doi.org/10.1613/jair.1523" />
	</analytic>
	<monogr>
		<title level="j">J. Artif. Intell. Res</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="page" from="457" to="479" />
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b119">
	<analytic>
		<title level="a" type="main">Efficient and effective spam filtering and re-ranking for large web datasets</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">V</forename><surname>Cormack</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">D</forename><surname>Smucker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">L A</forename><surname>Clarke</surname></persName>
		</author>
		<idno type="DOI">10.1007/s10791-011-9162-z</idno>
		<ptr target="https://doi.org/10.1007/s10791-011-9162-z" />
	</analytic>
	<monogr>
		<title level="j">Inf. Retr</title>
		<imprint>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="page" from="441" to="465" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b120">
	<analytic>
		<title level="a" type="main">Latent dirichlet allocation</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">M</forename><surname>Blei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">Y</forename><surname>Ng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">I</forename><surname>Jordan</surname></persName>
		</author>
		<ptr target="http://jmlr.org/papers/v3/blei03a.html" />
	</analytic>
	<monogr>
		<title level="j">J. Mach. Learn. Res</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="993" to="1022" />
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b121">
	<analytic>
		<title level="a" type="main">Random decision forests</title>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">K</forename><surname>Ho</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICDAR.1995.598994</idno>
		<ptr target="https://doi.org/10.1109/ICDAR.1995.598994" />
	</analytic>
	<monogr>
		<title level="m">Third International Conference on Document Analysis and Recognition, ICDAR 1995</title>
				<meeting><address><addrLine>Montreal, Canada</addrLine></address></meeting>
		<imprint>
			<publisher>IEEE Computer Society</publisher>
			<date type="published" when="1995">August 14 -15, 1995. 1995</date>
			<biblScope unit="page" from="278" to="282" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b122">
	<analytic>
		<title level="a" type="main">Learning whom to trust with MACE</title>
		<author>
			<persName><forename type="first">D</forename><surname>Hovy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Berg-Kirkpatrick</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Vaswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Hovy</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/N13-1132" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HTL 2013), Association for Computational Linguistics</title>
				<meeting>the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HTL 2013), Association for Computational Linguistics<address><addrLine>Atlanta, Georgia</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="1120" to="1130" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b123">
	<analytic>
		<title level="a" type="main">Aramis at Touché 2022: Argument detection in pictures using machine learning</title>
		<author>
			<persName><forename type="first">J</forename><surname>Braker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Heinemann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Schreieder</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2022 Evaluation Labs, CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b124">
	<analytic>
		<title level="a" type="main">Distinguishing cartoons images from real-life images</title>
		<author>
			<persName><forename type="first">M</forename><surname>Zaid</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>George</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Al-Khafaji</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">International Journal of Advanced Research in Computer Science and Software Engineering</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="page" from="91" to="95" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b125">
	<analytic>
		<title level="a" type="main">Boromir at Touché 2022: Combining natural language processing and machine learning techniques for image retrieval for arguments</title>
		<author>
			<persName><forename type="first">T</forename><surname>Brummerloh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">L</forename><surname>Carnot</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Lange</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Pfänder</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes Papers of the CLEF 2022 Evaluation Labs, CEUR Workshop Proceedings</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b126">
	<analytic>
		<title level="a" type="main">A new ANEW: evaluation of a word list for sentiment analysis in microblogs</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">Å</forename><surname>Nielsen</surname></persName>
		</author>
		<idno>WS.org</idno>
		<ptr target="http://ceur-ws.org/Vol-718/paper_16.pdf" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the ESWC2011 Workshop on &apos;Making Sense of Microposts</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<editor>
			<persName><forename type="first">M</forename><surname>Rowe</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Stankovic</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Dadzie</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Hardey</surname></persName>
		</editor>
		<meeting>the ESWC2011 Workshop on &apos;Making Sense of Microposts</meeting>
		<imprint>
			<date type="published" when="2011">2011</date>
			<biblScope unit="volume">718</biblScope>
			<biblScope unit="page" from="93" to="98" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
