<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">DistilBERT-based Argumentation Retrieval for Answering Comparative Questions Notebook for the Touché Lab on Argument Retrieval at CLEF 2021</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Alaa</forename><surname>Alhamzeh</surname></persName>
							<email>alaa.alhamzeh@insa-lyon.fr</email>
							<affiliation key="aff0">
								<orgName type="institution">Universität Passau</orgName>
								<address>
									<addrLine>Innstraße 41</addrLine>
									<postCode>94032</postCode>
									<settlement>Passau</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">INSA de Lyon</orgName>
								<address>
									<addrLine>20 Avenue Albert Einstein</addrLine>
									<postCode>69100</postCode>
									<settlement>Villeurbanne</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Mohamed</forename><surname>Bouhaouel</surname></persName>
							<email>mohamed.bouhaouel@uni-passau.de</email>
							<affiliation key="aff0">
								<orgName type="institution">Universität Passau</orgName>
								<address>
									<addrLine>Innstraße 41</addrLine>
									<postCode>94032</postCode>
									<settlement>Passau</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Előd</forename><surname>Egyed-Zsigmond</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">INSA de Lyon</orgName>
								<address>
									<addrLine>20 Avenue Albert Einstein</addrLine>
									<postCode>69100</postCode>
									<settlement>Villeurbanne</settlement>
									<country key="FR">France</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Jelena</forename><surname>Mitrović</surname></persName>
							<email>jelena.mitrovic@uni-passau.de</email>
							<affiliation key="aff0">
								<orgName type="institution">Universität Passau</orgName>
								<address>
									<addrLine>Innstraße 41</addrLine>
									<postCode>94032</postCode>
									<settlement>Passau</settlement>
									<country key="DE">Germany</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">DistilBERT-based Argumentation Retrieval for Answering Comparative Questions Notebook for the Touché Lab on Argument Retrieval at CLEF 2021</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">F41B078ABD1EADA333A6A8BEB896B9BE</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T20:43+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Comparative Question Answering</term>
					<term>Argumentation Mining</term>
					<term>Transfer Learning</term>
					<term>Argument Identification</term>
					<term>Computational Linguistic</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>In the current world, individuals are faced with decision making problems and opinion formation processes on a daily basis. For example, debating or choosing between two similar products. However, answering a comparative question by retrieving documents based only on traditional measures (such as TF-IDF and BM25) does not always satisfy the need. Thus, introducing the argumentation aspect in the information retrieval procedure recently gained significant attention. In this paper, we present our participation at the CLEF 2021 Touché Lab for the second shared task, which tackles answering comparative questions based on arguments. Therefore, we propose a novel multi-layer architecture where the argument extraction task is considered as the main engine. Our approach therefore is a pipeline of query expansion, argument identification based on DistilBert model, and sorting the documents by a combination of different ranking criteria.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Argumentation is a fundamental aspect of human communication and decision making. It can be defined as the logical reasoning humans use to derive a conclusion or justify their perspectives on a specific topic.</p><p>An argument consists of two elementary components: one or more premises and one claim (conclusion) <ref type="bibr" target="#b0">[1]</ref>. Argument mining is the automatic identification or extraction of arguments from unlabeled input text such as news articles, scientific papers, legal documents, reviews, and debates. Hence, argument mining has been investigated in various applications such as assisted writing and legal counselling. Furthermore, the arguments can serve as the keystone of an intelligent web search engine for question answering. In this context, the Webis Group 1 organized an argumentation retrieval event "Touché Lab at CLEF 2021" 2 that consists of two independent shared tasks: 1) Argument Retrieval for Controversial Questions and 2) Argument Retrieval for Comparative Questions. We present in this paper our adopted approach for participation as "Rayla Team" in the latter shared task. The main objective of this task is to support users facing some choice problem in their daily life. Given a comparative question (for instance, "Which is better, a Mac or a PC?"), the goal is to retrieve documents from the ClueWeb12 corpus <ref type="foot" target="#foot_0">3</ref> , and rank them based on different criteria, mainly, the arguments they provide.</p><p>In order to have more granularity control, our proposed architecture incorporates several units, each one performs a specific sub-task, namely: query expansion, argument extraction, scoring, and sorting.</p><p>The Touché organizers gave the option to use their own tool, TARGER <ref type="bibr" target="#b1">[2]</ref>, for the argument retrieval sub-task by providing a simple REST API to interact with. However, since this is the main block in our proposed pipeline, we decided to develop our own module based on the latest developments in the field of deep learning. Therefore, we implemented a new transfer learning model for Argument Identification based on DistilBert <ref type="bibr" target="#b2">[3]</ref>, a distilled version of BERT (Bidirectional Encoder Representations from Transformers) <ref type="bibr" target="#b3">[4]</ref>: smaller, faster, and lighter, with 97% of its language understanding capabilities <ref type="bibr" target="#b2">[3]</ref>.</p><p>The remaining of the paper is organized as follows: In Section 2, we go through a conceptual background of argument mining and the recent usage of transfer learning models towards it . In Section 3, we present our retrieving system architecture as well as the function of each unit. We discuss the proposed approach in Section 5. Finally, we summarize the main achievements in Section 6.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related work</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Argument Retrieval</head><p>Argumentative text identification is generally the first step of a complete argument mining system, then comes the argument components detection, and finally, the argument structure identification that considers the support and attack relations. State-of-the-art literature reports mainly classical machine learning models and very few attempts to use deep learning models (e.g. <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b6">7]</ref>).</p><p>In a classical machine learning model, the training and testing data are used for the same task and follow the same distribution. However, transfer learning aims to transfer previous learned knowledge from one source task (or domain) to a different target one, considering that source and target tasks and domains may be the same or may be different but related <ref type="bibr" target="#b7">[8]</ref>.</p><p>This can be useful in the argument mining domain from different perspectives. First of all, common knowledge about the language is obviously appreciated. Second, transfer learning can solve or at least help to solve one of the biggest challenges in the argument mining field, the lack of labeled datasets. Third, even available datasets are often of small size and very domain and task dependent. They may follow different annotations, argument schemes, and various feature spaces. This means that in each potential application of argument mining, we need argument experts to label a significant amount of data for the task at hand, which is definitely an expensive work in terms of time and human-effort. Hence, transfer learning will fine-tune pre-trained knowledge on a big dataset to serve another problem, and that's why we built on it for the argument identification stage.</p><p>Looking at the literature, we found only two studies addressing transfer learning models for argumentation tasks published in 2020. The first one is towards discriminating evidence related to Argumentation Schemes <ref type="bibr" target="#b8">[9]</ref> where the authors train classifiers on the sentence embeddings extracted from different pre-trained transformers. The second one is by Wambsganss et al. <ref type="bibr" target="#b4">[5]</ref> where the authors proposed an approach for argument identification using BERT <ref type="bibr" target="#b3">[4]</ref>. Our transfer learning model is based on a distilled version of BERT, proposed by <ref type="bibr" target="#b2">[3]</ref>, which retains 97% of the language understanding capabilities of the base BERT model, with a 40% less in size, and being 60% faster.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Comparative Question Answering</head><p>Question answering is a sophisticated form of information retrieval that started to develop as a field decades ago. However, with the growth of world wide web data, as well as of private databases, the need for more precise answers, well-expressed and shortly formulated is growing, too. Hence, several studies are devoted to the representation of natural language stated in the query and in the documents.</p><p>Extracting the arguments stated in the document is one way to clearly capture the grounded statements (premises) and the final conclusion (claim) presented in the text. Therefore, many recent works focus on the arguments as a potential tool for improving comparative question answering <ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b10">11]</ref>. Improving a retrieval model for argument-based comparison systems is the target of the Touché's Task 2, which has been addressed in 2020 <ref type="bibr" target="#b11">[12]</ref> as well.</p><p>The best result of last year was introduced by the team "Bilbo Baggins" of Abye et al. <ref type="bibr" target="#b12">[13]</ref>. They used a two-step approach: 1) query expansion by antonyms and synonyms, and 2) document ranking based on three measures: relevance, credibility, and support features. An overall score is deduced by summing up those three scores after weights multiplication. Eventually this approach can be improved by keeping the same pipeline and enhance the inner sub-tasks. For instance, instead of using a static classifier 'XGBoost' for the comparative sentence classification, a BERT-based model can ameliorate the robustness of the classifier and extend its operating range. In <ref type="bibr" target="#b13">[14]</ref>, Johannes Huck participated with a simple approach, the original topic is used as a single query to be sent to ChatNoir, and retrieve the top 20 documents. After documents retrieval, the author uses the TARGER API <ref type="bibr" target="#b1">[2]</ref> in order to extract arguments from each document, claims and premises are treated as arguments without any distinction. For the re-ranking, the BM25 algorithm has been used to determine the relevance of the extracted arguments with respect to the original query <ref type="bibr" target="#b13">[14]</ref>. Despite the simplicity of this implementation and the lack of introducing several measures in the ranking process, this approach could be integrated as one component in a bigger architecture in order to add an extra measure that may improve the stability of the overall system.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Approach and Implementation</head><p>In this section, we present our proposed approach and adopted methods to build a search engine for answering comparative questions based on argumentation identification. The overall architecture of our approach is presented in Figure <ref type="figure" target="#fig_0">1</ref>. It composes of a sequence of seven stages. We go through them individually in the upcoming sections. We used the same architecture to submit four runs with different configurations via TIRA platform <ref type="bibr" target="#b14">[15]</ref>. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Query Expansion</head><note type="other">Topic</note></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Query Expansion</head><p>Query Expansion (QE) is the process of reformulating a given query to improve retrieval performance and increase the recall of the candidate retrieval (i.e. to match additional documents). Our query expansion module involves a variety of techniques in order to generate three different queries to be passed to the next step, as the following:</p><p>• Query 1: is the original query itself.</p><p>• Query 2: is generated from the original query by: (1) removing English stop words, punctuation marks, and comparison adjectives or also called comparison operators. (2) Stemming of the remaining words to their base forms, and aggregating them together with conjunctive AND operator.</p><p>• Query 3: will be generated from the original query only if the latter contains a comparison operator, as follows:</p><p>-Search for synonyms and/or antonyms of the comparison operator of the query to get what is called the context of the comparison operator, whose size is 5 synonyms/antonyms in our case. -Remove English stop words and punctuation.</p><p>-Eliminate the comparison operator from the original query and transform the remaining words/terms to their base form. -We create 5 queries out of the original query (after the pre-specified changes: removing English stop words, punctuation and comparison operators), and each time adding one of the 5 synonyms/antonyms from the context of the comparison operator.</p><p>Those 5 output queries are sent to ChatNoir as one disjunctive OR-query: Query 3.</p><p>Comparison adjectives identification and words stemming are done automatically using using spaCy <ref type="foot" target="#foot_1">4</ref> . Synonyms and antonyms are hard-coded in the software due to the limited number of comparative operators in the topics of this year and the last year (10 adjectives in total). In Table <ref type="table" target="#tab_0">1</ref> we show an example of the query expansion output for the Topic 61 from Touché 2021 topics. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Retrieving the Documents by ChatNoir API</head><p>The ClueWeb12 document dataset for this task is easily accessible through the ChatNoir API<ref type="foot" target="#foot_2">5</ref>  <ref type="bibr" target="#b15">[16,</ref><ref type="bibr" target="#b16">17]</ref> that is based on the BM25F ranking algorithm. The BM25, also known as Best Matching 25 is a ranking method used by several search engines to determine the relevance of documents with regards to a query. It is based on the probabilistic retrieval framework and it is considered as the state-of-the-art among TF-IDF-like algorithms in information retrieval. BM25F is just a newer version of BM25 that takes the structure of documents into consideration.</p><p>The API takes a query as input, in our case we retrieve 40 documents for each of the 3 queries. The interaction with this API can be done using either GET or POST requests, several parameters could be included in the request to specify the corpus to retrieve from, and the type of returned information. It supports also a set of standard operators from web search services, and queries can be concatenated using Boolean operators, -, ". . . ", site:. . . etc.. In our approach, we used only POST requests, AND and OR operators to concatenate strings in Query 2 and Query 3.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Documents Aggregation</head><p>As Query 3 may be empty, we have a minimum of 80 and a maximum of 120 retrieved documents. Every document has a unique id (uuid), which we use to remove the redundant documents returned by more than one query. As a result we get N documents. One important remark at this phase, is that when removing a redundant document, we preserve its importance by including its retrieved scores (from ChatNoir) to the scores of the remaining unique document (having a unique uuid). This is done by summing up individual scores to obtain a single score for each set of redundant documents.</p><p>Initially, the ChatNoir API does not respond with the full content of the documents. Instead, it returns some metadata, including the unique uuid. To get a document full content, a new query is sent to ChatNoir to request the full HTML source page. Later on, we need to reveal only the main textual content from the HTML document. We achieve that by removing tags, advertisements and other cleaning processes. For that end, we used two HTML cleaning libraries: Boilerpy3<ref type="foot" target="#foot_3">6</ref> and Trafilatura <ref type="foot" target="#foot_4">7</ref> .</p><p>Both libraries are powerful and able to perform the main text extraction most of the time. However, the output of the two libraries is not always the same. In some cases, Trafilatura performs better than Boilerpy3 and extracts more text content, while the latter outperforms in other cases/documents. For this reason, we are using both libraries to extract the HTML content, the results are split into sentences and we take the union of those sentences in order to cover the entire main content of the document. The final list of sentences is sent to the Argument Extraction module, where each document is presented as a set of sentences.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.4.">Argument Extraction</head><p>In our particular task, we seek to detect the comparative sentences in the document, therefore, argument identification can be sufficient. Hence, we take the sentences from the document aggregation step and apply binary classification using the model presented in Figure <ref type="figure" target="#fig_1">2</ref> to label every sentence as an argument or Non-argument. Therefore, we searched for datasets that contain both argument and non-argument labels. Student Essays <ref type="bibr" target="#b17">[18]</ref> and Web discourse <ref type="bibr" target="#b18">[19]</ref> are public, very common used corpora which well satisfy this purpose.</p><p>The Student Essays corpus contains 402 Essays about 8 controversial topics. The annotation covers the following argument components: 'major claim', 'claim' and 'premise'. In addition, it shows the support/attack relations between them. Thus, it could be used in several argument mining tasks.</p><p>The User-generated Web Discourse corpus is a smaller dataset that contains 340 documents about 6 controversial topics in education such as homeschooling. The document may refer to an article, blog post, comment, or forum posts. This dataset is considered noisy, unrestricted and less formalized. The annotation has been done by <ref type="bibr" target="#b18">[19]</ref> according to Toulmin's model <ref type="bibr" target="#b19">[20]</ref>. Thus, it covers the following argument components: 'claim', 'premise', 'backing' and 'rebuttal'.</p><p>In order to deduct binary-labeled unified data for both corpora, we label any argument component (premise or claim) as an 'argument', and the rest of the text sentences as 'nonargument'.</p><p>The choice of BERT-based model is justified by the fact that among different existent transformers, BERT <ref type="bibr" target="#b3">[4]</ref> has recently gained a lot of attention. It seems to achieve the state of the art results in several NLP tasks <ref type="bibr" target="#b4">[5,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b6">7,</ref><ref type="bibr" target="#b20">21]</ref>.</p><p>For our particular task, we performed different experiments using several BERT-based models (BERT base, RoBERTa-base <ref type="bibr" target="#b21">[22]</ref>, DistilRoBERTa , DistilBERT) <ref type="foot" target="#foot_5">8</ref> and we achieved very similar results. Hence, we decided to use the DistilBERT model given that it is 40% less than BERT in size with a relevant in-line performance and a faster training/ testing time <ref type="bibr" target="#b2">[3]</ref>.    <ref type="formula">3</ref>) It adds the necessary padding to preserve the unique size for all sentences (we set 128 as a maximum length). The second block is the DistilBERT fine-tuned model that outputs mainly a vector of length of 768 (default length). In order to adapt the output of this pre-trained model to our specific task, we add a third block, which is a linear layer applied on top of the DistilBERT transformer, and outputs a vector of size 2. In this vector, the index of the maximum value reflects the predicted class id (Or simply it is the 𝑎𝑟𝑔𝑚𝑎𝑥 of the output vector). We trained the model for 3 epochs, using AdamW <ref type="bibr" target="#b22">[23]</ref> as an optimizer and Cross Entropy for the loss calculation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.5.">Scoring</head><p>This step is essential for any search engine system because users usually check out the first result links. Therefore, our objective now is to estimate the best matching between the query and the candidate answers, in order to sort them at the final stage. For that end, we investigate different scores based on different aspects. First of all, the document relevance which can be checked by ChatNoir BM25 score and Query hit count. However, even if a document content is relevant to the query, it may be faked or biased, we inspect the credibility of the document itself by considering: Page Rank score as well as Spam rank score.</p><p>Moreover, as we built our retrieval system based on arguments, we takes into consideration the argument quality level by three different scores: argument support, query-argument similarity, and argument BM25 score.</p><p>The complete details of our ranking scores are:</p><p>1. ChatNoir score: Returned form ChatNoir API indicating BM25 measure. 2. Query hit count: Indicates how many times the document is retrieved by the three queries <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b2">3]</ref>. 3. Page Rank score: Returned form ChatNoir API measuring the importance of the source website pages. 4. Spam rank score: Returned form ChatNoir API indicating the probability of the website to be a spam. 5. Argument support score: Represents the ratio of argument sentences among all existent sentences in the document. 6. Similarity score: Evaluates the similarity of two sentences based on the context and English language understanding using the SentenceTransformer<ref type="foot" target="#foot_6">9</ref> library <ref type="bibr" target="#b23">[24]</ref>. We calculate the similarity between the original query and every argumentative sentence in the document, and consider the average as the score. 7. Arg-BM25 score: Calculated on argumentative sentences of each document with respect to the original query. This is done through re-indexing the retrieved documents by creating new ones that contain only argumentative sentences. Then the arg-BM25 score of each document is calculated by querying the new argumentative documents with the original topic.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.6.">Normalization and Scores Combination</head><p>Furthermore, for the final score, we normalize all previously calculated scores, so that all values are between 0 and 1. These scores are summed up using particular weights. The setting up of the scoring weights is fixed manually based on the announced relevance judgments of last year </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Evaluation</head><p>The shown architecture in Figure <ref type="figure" target="#fig_0">1</ref> presents our base approach, from which we derive four submissions to the task-2 of Touché 2021 by modifying score weights and the number of retrieved and submitted documents. In Table <ref type="table" target="#tab_3">3</ref>, we present the used score weights for each run in addition to the number of submitted documents per topic, scores id are defined in Section 3.5. The results of each submission are conducted by the Touché committee through the manual labeling of the documents with the help of human assessors. For each run, two evaluation scores are calculated that defines the relevance and the rhetorical quality of the submitted documents. For the quality evaluation, our team Rayla comes out on top of the ranking list by achieving a score of nDCG@5= 0.675. While we are ranked third with respect to the relevance evaluation with an nDCG@5 score of 0.427. Table <ref type="table" target="#tab_4">4</ref> presents in more details the achievements of each run in more details. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Discussion</head><p>The proposed approach presents an advanced web search engine to answer comparative questions, each component in the architecture presented in Figure <ref type="figure" target="#fig_0">1</ref> plays an important role to achieve the best retrieval results. For instance, the query expansion component tries to make a first selection and build a set of topic-related documents. The three different queries generated from the original topic increase the coverage of the related documents and this works very well with the ChatNoir API since it is a basic BM25 retrieval system. Table <ref type="table" target="#tab_2">2</ref> presents the results of our DistilBERT model to accomplish the binary classification task (argument / non-argument) at the sentence level on different corpora. It shows that our DistilBERT model results are in-line to recent work of Wambsganss. et al. <ref type="bibr" target="#b4">[5]</ref>, it achieves very good results in argument classification with an accuracy of 0.8587 with respect to the merged corpora. Furthermore, the model is built on top of a text processing transformer, DistilBERT, and it is powered by English knowledge understanding, which makes the trained argument identification model scale on a wide variety of English topics and texts.</p><p>In order to re-rank the initially retrieved documents, as mentioned in Section 3.5, we included several measures in the final score calculation. Those measures define the importance of the document in terms of three criteria: (1) document relevance, (2) whether sufficient argumentative support is provided, and (3) trustworthiness and credibility. For the document relevance criteria, the ChatNoir score and the query-hit reflect well whether a document is relevant enough to the topic, while the spam rank score and the page rank score reveal the trustworthiness and credibility of the web page. For the argument support, three different scores are being considered, argument support score, argument similarity score and argument BM25 score.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>In this paper, we presented our participation at the shared task of argument retrieval for answering comparative questions. Our approach consists of a pipeline of several components for query expansion, retrieving the documents, extracting the arguments, and ranking them respectively. A main contribution is the transfer learning model we developed based on the DistilBert transformer, and using two different datasets for adapting it to the argument identification task. In addition, we consider different criteria to re-rank the output documents with respect to their relevance, the argument they contain, and their trustworthiness and credibility. In future work, we plan to use more advanced techniques for better query enrichment, in addition to using a machine learning model that learns the best weights of the scores mentioned in Section 3.5.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Global architecture of the submitted approach</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Transfer Learning Model Architecture</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 2</head><label>2</label><figDesc>Figure2shows our transfer learning model architecture to perform the argument identification task using DistilBERT. The first block is the Tokenizer that takes care of all the BERT input requirements, by doing the following: (1) It transforms the words of a sentence into an array of DistilBERT tokens. (2) It adds the special starting token ([CLS] token). (3) It adds the necessary padding to preserve the unique size for all sentences (we set 128 as a maximum length). The second block is the DistilBERT fine-tuned model that outputs mainly a vector of length of 768 (default length). In order to adapt the output of this pre-trained model to our specific task, we add a third block, which is a linear layer applied on top of the DistilBERT transformer, and outputs a vector of size 2. In this vector, the index of the maximum value reflects the predicted class id (Or simply it is the 𝑎𝑟𝑔𝑚𝑎𝑥 of the output vector). We trained the model for 3 epochs, using AdamW<ref type="bibr" target="#b22">[23]</ref> as an optimizer and Cross Entropy for the loss calculation.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Example of query expansion</figDesc><table><row><cell cols="2">Query id Generated query</cell></row><row><cell>Query 1</cell><cell>Who is stronger, Hulk or Superman?</cell></row><row><cell>Query 2</cell><cell>Hulk AND Superman</cell></row><row><cell>Query 3</cell><cell></cell></row></table><note>Hulk AND Superman AND strong OR Hulk AND Superman AND capable OR Hulk AND Superman AND powerful OR Hulk AND Superman AND able OR Hulk AND Superman AND weak</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2</head><label>2</label><figDesc>DistilBERT results on different datasets</figDesc><table><row><cell>Dataset</cell><cell cols="4">Accuracy Precision Recall F1-score</cell></row><row><cell>Student Essays</cell><cell>0.8727</cell><cell>0.8016</cell><cell>0.7477</cell><cell>0.7697</cell></row><row><cell>Web Discourse</cell><cell>0.7799</cell><cell>0.7718</cell><cell>0.6484</cell><cell>0.6655</cell></row><row><cell>Merged Corpora</cell><cell>0.8587</cell><cell>0.7887</cell><cell>0.7529</cell><cell>0.7683</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 3</head><label>3</label><figDesc>Configurations of each runTouche task 2. Thereby, we have done several experiments on the topics from last year, while changing each time the values of weights. Then, we took the best values of weights and applied them in different runs for this year submission. It this step, the documents are sorted based on the final score to get the top 20 documents that are highly relevant to answer the comparative topics. The final output is inserted into a text file while respecting the standard TREC format proposed by the Touché organization, so that each line should contain the following fields:</figDesc><table><row><cell>Score Weights</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 4</head><label>4</label><figDesc>Achievements of each run</figDesc><table><row><cell>Relevance</cell><cell>Quality</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_0">https://lemurproject.org/clueweb12/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_1">https://spacy.io/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_2">https://www.chatnoir.eu/doc/api/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_3">https://github.com/jmriebold/BoilerPy3</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_4">https://github.com/adbar/trafilatura</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_5">We used Transformers from huggingface.com for our experiments.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_6">https://github.com/UKPLab/sentence-transformers</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">A practical study of argument</title>
		<author>
			<persName><forename type="first">T</forename><surname>Govier</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2010">2010</date>
			<publisher>Cengage Learning</publisher>
			<pubPlace>Belmont, CA</pubPlace>
		</imprint>
	</monogr>
	<note>7th ed. ed.</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Targer: Neural argument mining at your fingertips</title>
		<author>
			<persName><forename type="first">A</forename><surname>Chernodub</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Oliynyk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Heidenreich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bondarenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Biemann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics: System Demonstrations</title>
				<meeting>the 57th Annual Meeting of the Association for Computational Linguistics: System Demonstrations</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="195" to="200" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<author>
			<persName><forename type="first">V</forename><surname>Sanh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Debut</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Chaumond</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Wolf</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1910.01108</idno>
		<title level="m">Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-W</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1810.04805</idno>
		<title level="m">Bert: Pre-training of deep bidirectional transformers for language understanding</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Unlocking transfer learning in argumentation mining: A domain-independent modelling approach</title>
		<author>
			<persName><forename type="first">T</forename><surname>Wambsganss</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Molyndris</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Söllner</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">15th International Conference on Wirtschaftsinformatik</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">N</forename><surname>Reimers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Schiller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Beck</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Daxenberger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Stab</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1906.09821</idno>
		<title level="m">Classification and clustering of arguments with contextualized word embeddings</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">Probing neural network comprehension of natural language arguments</title>
		<author>
			<persName><forename type="first">T</forename><surname>Niven</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H.-Y</forename><surname>Kao</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1907.07355</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">A survey on transfer learning</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">J</forename><surname>Pan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Yang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on knowledge and data engineering</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="page" from="1345" to="1359" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Transfer learning with sentence embeddings for argumentative evidence classication</title>
		<author>
			<persName><forename type="first">D</forename><surname>Liga</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Palmirani</surname></persName>
		</author>
		<ptr target="http://ceur-ws.org/Vol-2669/paper2.pdf" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 20th Workshop on Computational Models of Natural Argument co-located with the 8th International Conference on Computational Models of Argument (COMMA 2020)</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<editor>
			<persName><forename type="first">F</forename><surname>Grasso</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><forename type="middle">L</forename><surname>Green</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Schneider</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Wells</surname></persName>
		</editor>
		<meeting>the 20th Workshop on Computational Models of Natural Argument co-located with the 8th International Conference on Computational Models of Argument (COMMA 2020)<address><addrLine>Perugia, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020-09-08">September 8th, 2020. 2020</date>
			<biblScope unit="volume">2669</biblScope>
			<biblScope unit="page" from="11" to="20" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Answering comparative questions with arguments</title>
		<author>
			<persName><forename type="first">A</forename><surname>Bondarenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Beloucif</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Biemann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<idno type="DOI">10.1007/s13222-020-00346-8</idno>
	</analytic>
	<monogr>
		<title level="j">Datenbank-Spektrum</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="page" from="155" to="160" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bondarenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Franzek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Biemann</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1809.06152</idno>
		<title level="m">Categorizing comparative sentences</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Overview of Touché 2021: Argument Retrieval</title>
		<author>
			<persName><forename type="first">A</forename><surname>Bondarenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Gienapp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fröbe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Beloucif</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Ajjour</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Panchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Biemann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Wachsmuth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-72240-1_67</idno>
		<idno>doi:</idno>
		<ptr target="10.1007/978-3-030-72240-1\_67" />
	</analytic>
	<monogr>
		<title level="m">Advances in Information Retrieval. 43rd European Conference on IR Research (ECIR 2021)</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">D</forename><surname>Hiemstra</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M.-F</forename><surname>Moens</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Mothe</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Perego</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Sebastiani</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin Heidelberg New York</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2021">2021</date>
			<biblScope unit="volume">12036</biblScope>
			<biblScope unit="page" from="574" to="582" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">An open-domain web search engine for answering comparative questions</title>
		<author>
			<persName><forename type="first">T</forename><surname>Abye</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Sager</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">J</forename><surname>Triebel</surname></persName>
		</author>
		<ptr target="http://ceur-ws.org/Vol-2696/paper_130.pdf" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2020 -Conference and Labs of the Evaluation Forum</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<editor>
			<persName><forename type="first">L</forename><surname>Cappellato</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Eickhoff</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Névéol</surname></persName>
		</editor>
		<meeting><address><addrLine>Thessaloniki, Greece</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">September 22-25, 2020. 2696. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Development of a search engine to answer comparative queries</title>
		<author>
			<persName><forename type="first">J</forename><surname>Huck</surname></persName>
		</author>
		<ptr target="http://ceur-ws.org/Vol-2696/paper_178.pdf" />
	</analytic>
	<monogr>
		<title level="m">Working Notes of CLEF 2020 -Conference and Labs of the Evaluation Forum</title>
		<title level="s">CEUR Workshop Proceedings</title>
		<editor>
			<persName><forename type="first">L</forename><surname>Cappellato</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Eickhoff</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Névéol</surname></persName>
		</editor>
		<meeting><address><addrLine>Thessaloniki, Greece</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">September 22-25, 2020. 2696. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">TIRA Integrated Research Architecture</title>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Gollub</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Wiegmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-22948-1_5</idno>
	</analytic>
	<monogr>
		<title level="m">Information Retrieval Evaluation in a Changing World, The Information Retrieval Series</title>
				<editor>
			<persName><forename type="first">N</forename><surname>Ferro</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Peters</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin Heidelberg New York</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Elastic ChatNoir: Search Engine for the ClueWeb and the Common Crawl</title>
		<author>
			<persName><forename type="first">J</forename><surname>Bevendorff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Information Retrieval. 40th European Conference on IR Research (ECIR 2018)</title>
		<title level="s">Lecture Notes in Computer Science</title>
		<editor>
			<persName><forename type="first">L</forename><surname>Azzopardi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Hanbury</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Pasi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Piwowarski</surname></persName>
		</editor>
		<meeting><address><addrLine>Berlin Heidelberg New York</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Chat-Noir: A Search Engine for the ClueWeb09 Corpus</title>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Graßegger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Michel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Tippmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Welsch</surname></persName>
		</author>
		<idno type="DOI">10.1145/2348283.2348429</idno>
	</analytic>
	<monogr>
		<title level="m">35th International ACM Conference on Research and Development in Information Retrieval (SIGIR 2012)</title>
				<editor>
			<persName><forename type="first">B</forename><surname>Hersh</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Callan</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>Maarek</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Sanderson</surname></persName>
		</editor>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="1004">2012. 1004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Annotating argument components and relations in persuasive essays</title>
		<author>
			<persName><forename type="first">C</forename><surname>Stab</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of COLING 2014, the 25th international conference on computational linguistics: Technical papers</title>
				<meeting>COLING 2014, the 25th international conference on computational linguistics: Technical papers</meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="1501" to="1510" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Argumentation mining in user-generated web discourse</title>
		<author>
			<persName><forename type="first">I</forename><surname>Habernal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computational Linguistics</title>
		<imprint>
			<biblScope unit="volume">43</biblScope>
			<biblScope unit="page" from="125" to="179" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">E</forename><surname>Toulmin</surname></persName>
		</author>
		<idno type="DOI">10.1017/CBO9780511840005</idno>
		<title level="m">The Uses of Argument</title>
				<imprint>
			<publisher>Cambridge University Press</publisher>
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
	<note>2 ed</note>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Offended, Don&apos;t Be Abusive! Implicit/Explicit Messages in Offensive and Abusive Language</title>
		<author>
			<persName><forename type="first">T</forename><surname>Caselli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Basile</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Mitrović</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Kartoziya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Granitzer</surname></persName>
		</author>
		<author>
			<persName><surname>Feel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of LREC</title>
				<meeting>LREC</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ott</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Joshi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Levy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Stoyanov</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1907.11692</idno>
		<title level="m">Roberta: A robustly optimized bert pretraining approach</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Decoupled weight decay regularization</title>
		<author>
			<persName><forename type="first">I</forename><surname>Loshchilov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Hutter</surname></persName>
		</author>
		<ptr target="https://openreview.net/forum?id=Bkg6RiCqY7" />
	</analytic>
	<monogr>
		<title level="m">7th International Conference on Learning Representations, ICLR 2019</title>
				<meeting><address><addrLine>New Orleans, LA, USA</addrLine></address></meeting>
		<imprint>
			<publisher>OpenReview</publisher>
			<date type="published" when="2019">May 6-9, 2019. 2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Sentence-bert: Sentence embeddings using siamese bert-networks</title>
		<author>
			<persName><forename type="first">N</forename><surname>Reimers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</author>
		<ptr target="https://arxiv.org/abs/1908.10084" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</title>
				<meeting>the 2019 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
