=Paper=
{{Paper
|id=Vol-2696/paper_174
|storemode=property
|title=Ranking Arguments by Combining Claim Similarity and Argument Quality Dimensions
|pdfUrl=https://ceur-ws.org/Vol-2696/paper_174.pdf
|volume=Vol-2696
|authors=Lorik Dumani,Ralf Schenkel
|dblpUrl=https://dblp.org/rec/conf/clef/DumaniS20
}}
==Ranking Arguments by Combining Claim Similarity and Argument Quality Dimensions==
Ranking Arguments by Combining Claim Similarity and Argument Quality Dimensions Notebook for Touché: Argument Retrieval at CLEF 2020 Lorik Dumani and Ralf Schenkel Trier University, Germany {dumani,schenkel}@uni-trier.de Abstract In this paper we describe our submissions to the CLEF lab Touché, which addresses argument retrieval from a focused debate collection. Our ap- proach consists of a two-step retrieval. Step one finds the most similar claims to a query. Step two ranks the directly tied premises by the count of their convinc- ingness compared to other relevant premises, for which we aggregate the sum of three main argument quality dimensions. The final ranking consists of the product of the two components which are expressed as probabilities. 1 Introduction Argumentation is required not only in political debates, where people try to convince others of certain standpoints, e.g., political views. They are also essential for personal decision making, e.g., which smartphone to buy. Since the emergence of well-equipped computers and the increasingly sophisticated NLP methods, computational argumenta- tion has become a very popular field of research and seeks to help people to find good and strong arguments for their needs. In line with existing work, an argument is defined as a claim supported or attacked by at least one premise [13]. The claim is usually a con- troversial standpoint that should not be believed by a reader without further evidence (in form of premises). Touché [5,4] is the first lab on Argument Retrieval.1 It follows the classical TREC- style2 evaluation methodology and features two subtasks: 1. Argument retrieval from a focused debate collection to support argumentative con- versations by providing justifications for the claims. 2. Argument retrieval from a generic Web crawl to answer comparative questions with argumentative results and to support decision making. We participated in Task 1 and in this paper we provide a description of the implementa- tion of our approach. Our submissions to the task were done under the team name Don Quixote. Copyright c 2020 for this paper by its authors. Use permitted under Creative Commons Li- cense Attribution 4.0 International (CC BY 4.0). CLEF 2020, 22-25 September 2020, Thessa- loniki, Greece. 1 https://events.webis.de/touche-20/. 2 https://trec.nist.gov/tracks.html. Task 1 aims at supporting users directly and finding arguments, e.g., to strengthen their stance or to form opinions about certain issues by “strong” arguments. Thus, be- sides general relevance of the argument to the topic, argument quality dimensions (see the work of Wachsmuth et al. [15] for a survey of work on argument quality) will be also evaluated. From the top-k results of all submissions, pools of answers will be formed that will be assessed by crowd-sourcing; evaluation of the submissions will then be done with nDCG [10]. The retrieved arguments will be evaluated by the following qualities: (1) whether an argumentative text is logically cogent, (2) whether it is rhetorically well-written, and (3) whether it contributes to the users’ stance-building process (called utility). The Web provides innumerable documents that contain arguments. Especially in online portals controversial topics are discussed. Since basically anyone can partici- pate in the discussion, we can assume that practically all relevant aspects are addressed there. Furthermore, as most of the participants are not experts, the arguments might be written in an understandable language. Hence, the official data basis is the dataset from Ajjour et al. [1,2], which consists of controversial discussions from debate portals3 and ChangeMyView, and on which the argument search engine args.me [14] also bases its arguments. The lab’s participants can choose between the downloadable corpus and args’ API.4 In our implementation we work with the downloadable dataset. Moreover, the lab provides 50 topics on different areas such as “abortion” or “gay marriage”, for which the lab participants have to find strong arguments from the provided dataset. 2 Our Approach: Concept and Implementation In this section we give a short overview of our approach. First, we introduce the general concept, then we discuss the preprocessing of the provided data as well as another dataset that we use to estimate the convincingness of premises. Finally, we show how we find “strong” arguments. 2.1 Concept We follow the principles developed in [6,8], which we summarize here briefly. First the set of claims C = {c1 , c2 , . . . } of the collection is clustered such that claims with the same meaning are assigned to the same claim cluster, yielding Γ = {γ1 , γ2 , . . .} with γi , γj ⊆ C and γi ∩ γj = ∅ in an offline operation (see Section 2.2). Note that the claims in args.me are sometimes formulated as questions or topic titles. However, in the following we will only refer to claims. We precluster the claims because often, the same claim appears in different formulations in a large collection, and we want to consider all variants of the same claim at the same time. This is even more important for 3 The arguments were extracted from the following debate portals: debatewise.org, idebate.org, debatepedia.org, and debate.org. 4 args’ API: https://www.args.me/api-en.html. The downloadable dataset of args: https://zenodo.org/record/3734893#.Xw24QCgzaUk. premises, since premises with the same meaning, but different formulations are often used in many claims, but we want to retrieve that premise only once. We thus cluster the set of premises P = {p1 , p2 , . . .} and a set Π = {π1 , π2 , . . .} with πi , πj ⊆ P and πi ∩ πj = ∅ of premise clusters is formed such that premises with the same meaning are put into the same premise cluster. Now, given a query claim q, we apply a two-step retrieval approach. In the first step, our approach locates the claims in C that are most similar to q, following the observation that the more similar a claim is to a query, the more relevant are the premises of that claim to the query [7]. In the second step, we locate the claim clusters containing these claims, collect all premises related to a claim of one of these claim clusters, and finally determine the premise clusters to which these premises belong. For the ranking of premise clusters, we now apply a probabilistic ranking method. Thus, our goal is to compute P (πj |q), that is, the probability that πj is chosen as supporting or attacking premise cluster for q. We can calculate P (πj |q) by iterating over P all premises in πj and aggregating their individual probabilities, i.e., P (πj |q) = p∈πj P (p|q), where P (p|q) is the probability that p is chosen as support or attack for q. P (p|q) is defined by combining the two aforementioned steps, i.e., P (p|q) = P (c|q) · P (p|c). Formally, P (c|q) denotes the probability that c is chosen as a similar claim to q. P (p|c) denotes the probability that p is chosen as support (or attack) for c. In our previous work [6] we estimated P (p|c) exclusively via frequencies of premises. For our submissions to Task 1, we use a different approach [8] that also takes some di- mensions of argument quality into account. We describe this approach in the following subsections. 2.2 Preprocessing of the Provided Dataset Since the provided dataset by Ajjour et al. [1,2] originally consists of arguments with two components, that is, one claim with exactly one premise, we initially grouped all premises by their textually equal claim. Afterwards this grouping will become important because we calculate the convincingness of a premise in comparison to other premises of the same claim. Then the contextualized embeddings of both the claims and the premises were de- rived by using Sentence-BERT (SBERT) [12].5 For the clustering of claims as well as premises we followed the approach of our prior work [6] and implemented an ag- glomerative clustering applying Euclidean distance, the average linkage method, and a dynamic tree cut [11].6 5 The framework used for this is available on https://github.com/UKPLab/ sentence-transformers. The model we used for calculating the embeddings is “roberta-large-nli-stsb-mean-tokens”, yielding embeddings of 1,024 dimen- sions each. 6 For the agglomerative clustering we used the scripting language R and the packages STATS and FAST C LUSTER. 2.3 Including another Dataset to Estimate the Convincingness of Premises Wachsmuth et al. [15] provide a dataset in which three experts assessed 320 arguments with respect to 15 argument quality dimensions. The arguments are distributed over 32 issue-stance pairs, i.e., 16 topics with two polarities and 10 premises each. Among these 15 dimensions there are the three main dimensions: (1) logical quality in terms of the cogency or strength of an argument, (2) rhetorical quality in terms of the persuasive effect of an argument or argumentation (called effectiveness), and (3) dialectical quality in terms of the reasonableness of argumentation for resolving issues. We considered the mean assessment values for the three main dimensions cogency, rea- sonableness and effectiveness and integrated the idea of Habernal and Gurevych [9] by deriving all combinations of (premise1 , premise2 ) pairs with premises from the same issue-stance pairs and labels “1” or “2”, whereby the labels signal which premise has a higher score with respect to a dimension. Pairs with equal mean value were omit- ted. Then, for the two premises of each pair as well as the corresponding (issue,stance) pair, we derived their SBERT embeddings, processed them to vectors consisting of the embeddings of the two premises each with the pointwise sum, difference, and product to the embedding of the corresponding (issue,stance) pair, yielding a vector of 6,144 dimension per (premise1 , premise2 ) pair.7 Then, we tested standard classifiers such as gradient boosting, logistic regression, or random forest with 32-fold-cross-validation and found that random forest performs best for cogency and effectiveness. For reason- ableness Logistic Regression performed only slightly better. Using these best classi- fiers per dimension, we were able to precalculate the dimension convincing frequencies (DCFs) of the premises in the datset by Ajjour et al. [1,2]. Here, the DCF of a premise of a claim is calculated as the count how often the premise was better than the other premises belonging to the same claim in a cross comparison. Now, both claims and premises could be indexed in two separate inverted indexes with the cluster and DCF information. We used Apache Lucene to build the indexes.8 2.4 Finding Strong Arguments For each of the 50 topics which we regard as queries, we started by finding the most similar claims (result claims) using Divergence from Randomness (DFR) [3], because our previous work [7] implies that DFR is well suited for this task. Then all premises belonging to claims that are in the same cluster as the result claims were localized. The set of premises was then expanded with the set of premises in the same premise clusters, yielding the set of result premises. Before calculating the premise cluster scores, first the premises were ranked indi- vidually. The ranking of these consists of the two components (1) similarity of the query 7 The input can be determined by the elementwise computed Cartesian product of the embed- dings of the following three sets, in compliance with the below order. The difference is positive. {premise1 , premise2 }, {+, −, ∗}, {(issue,stance)-pair}. 8 https://lucene.apache.org/. to the claim and (2) the sum of the three different DCFs per premise (see Section 2.3). Both (1) and (2) were normalized to have values between 0 and 1, allowing to use them like probabilities (P (p|c) in the description above). The cluster scores were determined by aggregating the scores of the individual premises of the same cluster. From each cluster, only one representative was selected; in our implementation this is the longest premise as we followed the intuition that a longer premise is also more specific and therefore may be better suited as a representative. As trec eval sorts documents by the score values and not by rank values, it is im- portant to handle tied scores. Furthermore, it is the score (integer or floating point) that is relevant for the TREC evaluation in the ranking. Therefore, the representatives were sorted in descending order by cluster score, then by length to break ties, and alphabeti- cally if also the length was the same. To reflect this in the ranking, of all representatives with the same initial score, the scores were increased by the smallest possible delta in Java (10−17 ) starting from the premise at the bottom. Subsequent Adjustions We manually reviewed the results of our retrieval at a cutoff value of 30 and found that premises with less than 30 characters were usually com- pletely useless as they are too unspecific or nonsense, so we removed them from the results. 3 Conclusion In this paper we outlined our contribution (team Don Quixote) to the CLEF lab Touché. First we cluster claims and premises in an offline operation by their meaning. For a given query, we then work with a two-step retrieval process that first finds all similar claims and then, using the clusters, finds the relevant premises. For the ranking, we then calculate (1) the similarity of claim and query, and (2) the frequency with which a premise is more convincing than other relevant premises with respect to the three main argument quality dimensions cogency, reasonableness, and effectiveness. Describing (1) and (2) as probabilities, a ranking can be generated via their product. The code will be made available shortly. Acknowledgements This work has been funded by the Deutsche Forschungsgemeinschaft (DFG) within the project ReCAP, Grant Number 375342983 - 2018-2020, as part of the Priority Program “Robust Argumentation Machines (RATIO)” (SPP-1999). References 1. Ajjour, Y., Wachsmuth, H., Kiesel, D., Riehmann, P., Fan, F., Castiglia, G., Adejoh, R., Fröhlich, B., Stein, B.: Visualization of the topic space of argument search results in args.me. In: Blanco, E., Lu, W. (eds.) EMNLP. pp. 60–65. Association for Computational Linguistics (2018). https://doi.org/10.18653/v1/d18-2011, https://doi.org/10.18653/v1/d18-2011 2. Ajjour, Y., Wachsmuth, H., Kiesel, J., Potthast, M., Hagen, M., Stein, B.: Data acquisition for argument search: The args.me corpus. In: Benzmüller, C., Stuckenschmidt, H. (eds.) KI. Lecture Notes in Computer Science, vol. 11793, pp. 48–59. Springer (2019). https://doi.org/10.1007/978-3-030-30179-8 4, https://doi.org/10.1007/978-3-030-30179-8\_4 3. Amati, G., van Rijsbergen, C.J.: Probabilistic models of information retrieval based on measuring the divergence from randomness. ACM Transactions on Information Systems 20(4), 357–389 (2002). https://doi.org/10.1145/582415.582416 4. Bondarenko, A., Fröbe, M., Beloucif, M., Gienapp, L., Ajjour, Y., Panchenko, A., Biemann, C., Stein, B., Wachsmuth, H., Potthast, M., Hagen, M.: Overview of Touché 2020: Argument Retrieval. In: Working Notes Papers of the CLEF 2020 Evaluation Labs (Sep 2020) 5. Bondarenko, A., Hagen, M., Potthast, M., Wachsmuth, H., Beloucif, M., Biemann, C., Panchenko, A., Stein, B.: Touché: First shared task on argument retrieval. In: ECIR. Lecture Notes in Computer Science, vol. 12036, pp. 517–523. Springer (2020). https://doi.org/10.1007/978-3-030-45442-5 67, https://doi.org/10.1007/978-3-030-45442-5\_67 6. Dumani, L., Neumann, P.J., Schenkel, R.: A framework for argument retrieval - ranking argument clusters by frequency and specificity. In: ECIR. Lecture Notes in Computer Science, vol. 12035, pp. 431–445. Springer (2020). https://doi.org/10.1007/978-3-030-45439-5 29, https://doi.org/10.1007/978-3-030-45439-5\_29 7. Dumani, L., Schenkel, R.: A systematic comparison of methods for finding good premises for claims. In: SIGIR. pp. 957–960 (2019), https://doi.org/10.1145/3331184.3331282 8. Dumani, L., Schenkel, R.: Quality-aware ranking of arguments. In: CIKM (2020), accepted 9. Habernal, I., Gurevych, I.: Which argument is more convincing? Analyzing and predicting convincingness of web arguments using bidirectional LSTM. In: ACL (2016), https://www.aclweb.org/anthology/P16-1150/ 10. Järvelin, K., Kekäläinen, J.: Cumulated gain-based evaluation of IR techniques. ACM Transactions on Information Systems 20(4), 422–446 (2002). https://doi.org/10.1145/582415.582418, http://doi.acm.org/10.1145/582415.582418 11. Langfelder, P., Zhang, B., Horvath, S.: Dynamic tree cut: In-depth description, tests and applications (2009), https://horvath.genetics.ucla.edu/html/ CoexpressionNetwork/BranchCutting/Supplement.pdf 12. Reimers, N., Schiller, B., Beck, T., Daxenberger, J., Stab, C., Gurevych, I.: Classification and clustering of arguments with contextualized word embeddings. In: ACL. pp. 567–578 (2019), https://www.aclweb.org/anthology/P19-1054/ 13. Stede, M., Afantenos, S.D., Peldszus, A., Asher, N., Perret, J.: Parallel discourse annotations on a corpus of short texts. In: LREC (2016), http: //www.lrec-conf.org/proceedings/lrec2016/summaries/477.html 14. Wachsmuth, H., Potthast, M., Khatib, K.A., Ajjour, Y., Puschmann, J., Qu, J., Dorsch, J., Morari, V., Bevendorff, J., Stein, B.: Building an argument search engine for the web. In: ArgMining@EMNLP. pp. 49–59. Association for Computational Linguistics (2017). https://doi.org/10.18653/v1/w17-5106, https://doi.org/10.18653/v1/w17-5106 15. Wachsmuth, H., Stein, B., Hirst, G., Prabhakaran, V., Bilu, Y., Hou, Y., Naderi, N., Alberdingk Thijm, T.: Computational argumentation quality assessment in natural language. In: EACL. pp. 176–187 (2017), https://aclweb.org/anthology/E17-1017/