=Paper=
{{Paper
|id=Vol-2696/paper_207
|storemode=property
|title=University of Amsterdam at CLEF 2020
|pdfUrl=https://ceur-ws.org/Vol-2696/paper_207.pdf
|volume=Vol-2696
|authors=Mahsa S. Shahshahani,Jaap Kamps
|dblpUrl=https://dblp.org/rec/conf/clef/ShahshahaniK20a
}}
==University of Amsterdam at CLEF 2020==
University of Amsterdam at CLEF 2020 Notebook for the Touché Lab on Argument Retrieval at CLEF 2020 Mahsa S. Shahshahani and Jaap Kamps University of Amsterdam {m.shahshahani,kamps}@uva.nl Abstract This paper documents the University of Amsterdam’s participation in CLEF 2020 Touché Track. This is the first year this track has been introduced at CLEF, and we were attracted to participate in it due to its potentialities for Parliamentary debates we are currently working on. This track consists of two tasks: Conversational Argument Retrieval and Comparative Argument Retrieval. We submitted a run to both tasks. For the first task, we used a combination of the traditional BM25 model and learning to rank models. BM25 model helps to retrieve relevant arguments, and learning to rank model helps to re-rank the list and put stronger arguments on top of the list. For the second task, Compara- tive Argument Retrieval, we proposed a pipeline to re-rank documents retrieved from Clueweb using three features: PageRank scores, web domains, and argu- mentativeness. Preliminary results on 5 queries have shown that this heuristic pipeline may help to achieve a balance among three important dimensions: rele- vance, trustworthiness, and argumentativeness. 1 Introduction We believe that we passed the era in which search engines were supposed to only give us a ranked list of documents or answers and they have more potentialities to help us in decision making process. Argument retrieval task has been defined to formulate this problem. Touché track at CLEF [3] offers an opportunity to work on this interesting problem having access to a debate corpus from two different points of view: Looking for different views about a problem in debates between opponents and supporters of a controversial issue, and looking for comparative opinions about different alternative. This track consists of a different task (two tasks in total) for each point of view and we submitted a run to both tasks. In this paper, we cover both tasks; first, we give a high-level summary of the first task, followed by our detailed approach. Then, we cover the same for the second task. Finally, we will conclude the paper by mentioning our main contributions and findings. We will add results section whenever the results would be out. Copyright c 2020 for this paper by its authors. Use permitted under Creative Commons Li- cense Attribution 4.0 International (CC BY 4.0). CLEF 2020, 22-25 September 2020, Thessa- loniki, Greece. 2 Conversational Argument Retrieval For detailed information about CLEF track’s experimental setup, we refer to the overview paper [4] and to the track homepage.1 However, we provide a high-level summary to make this paper self-contained. 2.1 Task definition The goal of this task is to retrieve relevant arguments from online debate portals, given a query on a controversial topic. Corpus Args.me [1] has been created by crawling arguments from 7 debate websites. It includes 387,606 arguments taken from 59,637 debates. A search engine based on Elasticsearch has been set to make it easier to work with this corpus [10]. This search engine ranks arguments using BM25 ranking algorithm. Different approaches can be used later to re-rank these retrieved arguments. Queries 50 controversial topics have been picked for this task. Each topic has both pro and con arguments in the corpus. Quality Assessment Proposed approaches are supposed to retrieve “strong” arguments. An argument is consiagstuhl ered strong if it is topically relevant, logically cogent, rhetorically well-written, and useful to help in stance-building process. Here, we define these assessment dimensions taken from [9] and the annotation guidelines for Dagstuhl- 15512 ArgQuality Corpus.2 Topical relevance: As every other ranking task, retrieved arguments should provide the user with relevant information about the query. Besides relevance, in general, there are three main dimensions for assessing the quality of arguments: logic, rhetoric, and dialectic. Logical cogency: An argument with acceptable premises that are relevant and sufficient to the argument’s conclusion is considered “cogent” [7]. Rhetorical well-writtenness: An argument is called “rhetorically well-written” if it is effective and successful in persuading a target audience of a conclusion [2]. Dialectic: An argument is considered reasonable if it contributes in the users’ stance- building process regards a given issue in a way that is acceptable to everyone. 2.2 Our Approach We treated this task as a re-ranking problem. In the absence of training data, we have to use unsupervised approaches. However, we used an existing debate dataset created for studying argument quality assessment to train a classifier. First, we describe this corpus. Later, we explain our approach in three consecutive steps. 1 https://events.webis.de/touche-20/shared-task-1.html 2 http://www.arguana.com Corpus Dagstuhl-15512 ArgQuality [9] includes 20 arguments for each of 16 queries. Three annotators have annotated these 320 arguments on 15 dimensions with three la- bels. However, we only use four dimensions: cogency, effectiveness, reasonableness, and overall quality. The set of labels includes ordinal scores from 1 (low) to 3 (high) for all dimensions. We used majority voting technique to get the label for each dimension, and substituted the ones labeled as 3 with 2 as there are a very few samples with label “3” in the corpus . Approach We created the final ranked list in three steps. Before explaining each step, we explain the way we represented arguments. Argument Representation: We used pre-trained BERT-base [6] model from Hugging- Face Transformers framework in python to represent arguments. As BERT model im- poses a limit on the length of documents after tokenization, we used the first 512 tokens in an argument if its length exceeds this limit. First step- Ranking: BM25 is the traditional unsupervised ranking model which scores the relevancy of documents (here arguments) in regards to queries based on the fre- quency of common terms between the query and argument. We ranked arguments for each topic based on BM25 using args.me search engine. Second step- Classification: We trained a classifier on Dagstuhl-15512 ArgQuality corpus to recognize and label cogency, well-writtenness, reasonableness and overall quality of each argument. Later, we applied this classifier to all retrieved arguments in the ranked list from the fist step. We got the majority voting for each dimension; in all of the arguments the majority for all four dimensions are the same. It is aligned with the conclusion in the main paper [9] which indicates that cogency, effectiveness, and reasonableness correlate strongly with overall quality, and also much with each other. We trained two classifiers on 90% of data: A decision tree classifier, and an SVM classifier. The accuracy on the remaining 10% of data has been shown in Table 1. We used scikit-learn framework for python 3 to train both classifiers. For decision tree, we used “gini” criterion, and set minimum required samples to split to 2. For SVM, we used “rbf” kernel, and set regularization parameter to 1. We selected SVM classifier for the further step. Third step-Re-ranking: In the third step, we re-ranked retrieved arguments from the first step using learning-to-rank models. We use the output of step 2 as a feature in step 3 to re-rank the arguments. We trained three different learning to rank models: Ranknet, RandomForests, and LambdaRank. We used RankLib3 library with default sets of parameters to apply these models. In order to train learning to rank models, we used argument representations based on BERT model, the output of the second step, and two additional features based on named entities. We defined two binary features in- dicating the presence of numerical named entities (percent, quantity, money) and other 3 https://github.com/codelibs/ranklib entities (person, location, organization) in the argument. We showed the number of ar- guments with and without these entities in Figure 1 and Figure 2. These figures show the difference in the distribution of each of these features in strong (label=2) and weak (label=1) arguments. Arguments using these kinds of entities are more likely providing users with persuasive and effective information to make their stance, and this can lead to a more probability to be labeled as “strong”. We trained learning to rank models on Dagstuhl dataset, and applied the best model (Ranknet) to re-rank retrieved arguments from args.me dataset for all 50 topics in the shared task. We trained models on 90% of data and reported accuracy on the remaining 10% of data in Table 2. Table 1. classifiers Model Accuracy Decision Tree 0.43 SVM 0.53 Figure 1. Having/lacking numerical named entities (percent, money, quantity) Table 2. LTR Model NDCG@1 NDCG@5 RankNet 0.873 0.9587 Random Forests 0.7333 0.9155 LambdaRank 0.7778 0.9291 Figure 2. Having/lacking other named entities (person, organization) 2.3 Results In the final relevance judgments, 30 documents for each topic have been annotated using 6 labels: -2,1,2,3,4 and 5. Classifier We ran our classifier that was trained on Dagstuhl-15512 ArgQuality corpus on arguments from judgments to see if the results are correlated with the final judgment. Unfortunately, we observed that the classifier does not work well and return label ’1’ for more than 90% of the judged arguments. This suggests that the classifier does not play a role in the final results of our learning to rank model. This is not surprising as the dataset we trained our SVM classifier on is very different from the test set. The arguments in the training set are very short and consist of only one to three sentences, while we are labeling a set of complete documents in the test set. However, as the BERT representation has a limit on the size of the input text, for the longer texts we only used the first 512 tokens of each argument. Using more advanced approaches such as averaging over sliding windows of 512 tokens might make the classifier useful. Entities Similar to Figure 1 and Figure 2, we looked into the distribution of numeri- cal and other types of entities in relevant and non-relevant documents. We considered documents with label ’-2’ as non-relevant, and other documents in the judgment file as relevant. Results have been shown in Figure 3 and 4. As it is obvious from the figures, the distribution of entities is the opposite of what we observed in the dataset we used at the time of developing our model. However, it still can be used as a feature as it shows a little difference between relevant and non-relevant documents. Query Length We looked into per query results to get an insight into the way our model works. We used NDCG@5 metric as it has been the main metric for the shared task. Figure 3. Having/lacking numerical named entities (percent, money, quantity) Figure 4. Having/lacking numerical named entities (person, organization) Table 3. Results-NDCG@10 metric Model NDCG@1 NDCG@5 NDCG@10 MAP UvATask1LTR 0.5214 0.5548 0.3709 0.1129 As we used the whole topic (without removing stop words), the model works better for shorter queries (Figure 5). Figure 5. Average Query Length for the different ranges of NDCG@5) 3 Comparative Argument Retrieval Similar to the previous section, for detailed information about this task’s experimental setup, we refer to the overview paper [4] and to the track homepage.4 However, we provide a high-level summary. 3.1 Task definition The goal of this task is to retrieve and rank documents from web that help to answer a comparative question from “everyday life”. Corpus Clueweb12 is a dataset created by crawling 733,019,372 web documents seeded with 2,820,500 urls from Clueweb09 [5]. We used a publicly available search engine [8] based on Elasticsearch to retrieve documents from Clueweb12 based on BM25 ranking model. Queries 50 comparative topics from everyday life have been picked for this task. 3.2 Our approach We treated this task as a re-ranking problem. In the absence of training data, we have to use unsupervised approaches. We labeled retrieved documents for 5 topics to have an insight of how our heuristic approach works. In the first step, we used ChatNoir search engine to rank documents retrieved for each topic. In the second step, we used three different features to re-rank them. Here, we introduce the features we used, followed by our heuristic approach to combine them and create the final ranked list. 4 https://events.webis.de/touche-20/shared-task-2.html Argumentativeness We trained a simple SVM classifier based on data from args.me corpus and Clueweb to distinguish between argumentative and non-argumentative doc- uments. Similar to the first task, we used BERT-based model from HuggingFace Trans- formers library for Python to represent documents, and we used the first 512 tokens in a document if its length exceeds the limit of BERT model. To train the classifier, we used a small sample from each corpus. These samples are created by submitting all 50 controversial queries in the first task to both corpora and got up to 100 documents for each query. Then, we manually removed argumen- tative documents from the sample taken from Clueweb and considered the remaining documents as negative examples. All retrieved documents from args.me corpus have been considered as positive examples. The final training set consists of 3000 positive and 3000 negative examples. Then, We trained a simple SVM classifier on 80% of the data, and evaluated it on the remaining 20% of documents. It achieved 87% in terms of accuracy. All parameters for the argumentativeness classifier have been set to their default values in Scikit-learn5 library for Python. Web domains Clueweb has been formed by crawling web documents with some post- filters. But, the goal of this task is to retrieve documents including personal opinions or suggestions. Thus, documents from particular domains like Wikipedia are not desirable. On the contrary, documents from discussion forums, debate websites, and blogs can be very helpful. Having this intuition in mind, we defined a binary feature that indicates if the source URL for a discussion contains ’forum’ or ’blog’ terms to give a bonus to web pages from discussion forums or blogs. PageRank Although desired documents are those from discussion forums and personal blogs, they should also be trustworthy. To take trustworthiness into account, we used page rank scores to prioritize documents taken from more reliable sources. In ChatNoir search engine, every returned document has been associated with a PageRank score. We directly used these returned scores. Re-ranking We introduced three features, and our goal is to re-rank documents based on a combination of these features (argumentativeness, domain addresses , and PageR- ank scores). To generate the final ranked list, we make a heuristic ranking pipeline in four steps: – The 1st step: The initial ranked list is taken from ChatNoir search engine. ChatNoir retrieves documents from Clueweb and ranks them using traditional BM25 ranking model. We use the whole topic title as the query. We also examined submitting the query after removing stop words or just using the entities in the topic title. But, using the whole topic title worked better. – The 2nd step: Page-rank scores are used to re-rank the list from the first step in descending order. This may result in putting a document, initially ranked very low, on top of the list. In order to avoid this, moving documents in the ranked list is limited to a maximum of 10 positions. 5 https://scikit-learn.org/stable/supervised_learning.html#supervised-learning Table 4. Preliminary Results- NDCG@10 metric Model Relevance Argumentativeness Trustworthiness initial 0.87 0.71 0.81 pagerank 0.89 0.66 0.87 domain 0.84 0.72 0.80 argumentative classifier 0.80 0.84 0.79 mixed 0.84 0.78 0.82 – The 3rd step: Web domain are used to re-rank ranked list from the second step. To do this, the documents with positive domain feature (which means the document is taken from a blog or discussion forum) are put on top of the list. This has been performed within every 10 documents in the list. We split the list into chunks of 10 documents and we keep their relative positions. – The 4th step:All retrieved documents in the ranked list from the third step are clas- sified using the argumentativeness classifier we have trained. Documents classified as positive are put on top of the list. We keep their relative positions. This has been performed within every chunk of 10 documents in the list. We put this limits on moving documents in the list with this intuition in mind that relevance should be prioritized in comparison with trustworthiness and argumentative- ness. 3.3 Results Since the final judgment file does not consist of separate judgment lists for three dif- ferent dimensions (relevance, trustworthiness, and argumentativeness), we include our preliminary results too. Preliminary Results: To gain an insight into the effectiveness of our heuristic model, we manually labeled 10 retrieved documents for 5 queries. We labeled documents using three labels: 0 for non-relevant, non-argumentative, or untrustworthy; 1 for relevant, ar- gumentative, or trustworthy, and 2 for highly-relevant, highly-argumentative, or highly trustworthy. We evaluated the top 10 documents for each ranked list: BM25, re-ranked by PageR- ank scores, re-ranked by web domains, re-ranked by argumentativeness, and mixed model. Evaluation results have been reported in Table 3. The heuristic mixed model does not achieve the same performance as BM25 in terms of relevance, the same performance as the argumentative classifier model in terms of argumentativeness, and the same per- formance as the PageRank model in terms of trustworthiness. But, it seems that it struck a balance among all three dimensions. Table 5. Final Results Model NDCG@1 NDCG@5 NDCG@10 initial 0.5068 0.4480 0.4196 pagerank 0.4723 0.4256 0.4087 domain 0.4100 0.3845 0.4132 argumentative classifier 0.4281 0.4123 0.4023 UvATask2SVM 0.5000 0.4464 0.4185 Final Results: The final results for every step of our model have been reported in Figure 5. Since three different aspects had been defined for evaluation of this task, We expected to receive three judgment sets. However, the judgment file issued only the overall relevance, which we sacrificed for the other two dimensions. Initially, we processed the data and implemented our models based on the three different aspects which had been defined previously. Those include relevance, trustwor- thiness, and argumentativeness. Given the fact that we received the overall relevance file only, we extracted our results purely based on the relevance factor for the evalu- ation of this task. This categorically implies that the obtained results could have been more optimized if we had performed our preliminary assessment based on one aspect only. 4 Conclusion This paper documents our first participation in the Touché 2020 Track. We explained our approaches for both tasks in the track: Conversational Argument Retrieval, and Comparative Argument Retrieval. For the Conversational Argument Retrieval task, we used an existing argument qual- ity assessment dataset to train a classifier and re-rank arguments based on the output of this classifier. We showed that named entities are important features to distinguish be- tween strong and weak arguments on the preliminary data. But, the final results show that neither classifier nor entities do not help. However, it is worth mentioning that the classifier had been trained on a completely different set of arguments. So, if we train it on the same data from args.me, it might be useful for the final results. This needs further investigations to be proved. For the Comparative Argument Retrieval task, we introduced three features to re- rank arguments taken from Clueweb. We proposed a pipeline to combine different as- pects (relevance, trustworthiness, and argumentativeness) to create the final ranked list. Preliminary results have shown that this heuristic pipeline may successfully strike a balance between all three dimensions. However, the final evaluation has been done only on relevance. This caused our method to be sub-optimal. We hope and expect that the valuable bench-marking data created at Touché track will be of great value to motivate, and greatly facilitate, further research into argument retrieval. Acknowledgments This research was supported in part by the Netherlands Organization for Scientific Research (NWO, grant # CISC.CC.016, ACCESS project). Views expressed in this paper are not necessarily shared or endorsed by those funding the research. References 1. Ajjour, Y., Wachsmuth, H., Kiesel, J., Potthast, M., Hagen, M., Stein, B.: Data acquisition for argument search: The args.me corpus. In: Benzmüller, C., Stuckenschmidt, H. (eds.) KI 2019: Advances in Artificial Intelligence - 42nd German Conference on AI, Kassel, Germany, September 23-26, 2019, Proceedings. Lecture Notes in Computer Science, vol. 11793, pp. 48–59. Springer (2019). https://doi.org/10.1007/978-3-030-30179-8_4, https://doi.org/10.1007/978-3-030-30179-8_4 2. Blair, J.A.: Groundwork in the theory of argumentation: Selected papers of J. Anthony Blair, vol. 21. Springer Science & Business Media (2012) 3. Bondarenko, A., Fröbe, M., Beloucif, M., Gienapp, L., Ajjour, Y., Panchenko, A., Biemann, C., Stein, B., Wachsmuth, H., Potthast, M., Hagen, M.: Overview of Touché 2020: Argument Retrieval. In: Working Notes Papers of the CLEF 2020 Evaluation Labs (Sep 2020) 4. Bondarenko, A., Hagen, M., Potthast, M., Wachsmuth, H., Beloucif, M., Biemann, C., Panchenko, A., Stein, B.: Touché: First shared task on argument retrieval. In: Jose, J.M., Yilmaz, E., Magalhães, J., Castells, P., Ferro, N., Silva, M.J., Martins, F. (eds.) Advances in Information Retrieval - 42nd European Conference on IR Research, ECIR 2020, Lisbon, Portugal, April 14-17, 2020, Proceedings, Part II. Lecture Notes in Computer Science, vol. 12036, pp. 517–523. Springer (2020). https://doi.org/10.1007/978-3-030-45442-5_67, https://doi.org/10.1007/978-3-030-45442-5_67 5. Clarke, C.L.A., Craswell, N., Soboroff, I.: Overview of the TREC 2009 web track. In: Voorhees, E.M., Buckland, L.P. (eds.) Proceedings of The Eighteenth Text REtrieval Conference, TREC 2009, Gaithersburg, Maryland, USA, November 17-20, 2009. NIST Special Publication, vol. 500-278. National Institute of Standards and Technology (NIST) (2009), http://trec.nist.gov/pubs/trec18/papers/WEB09.OVERVIEW.pdf 6. Devlin, J., Chang, M., Lee, K., Toutanova, K.: BERT: pre-training of deep bidirectional transformers for language understanding. In: Burstein, J., Doran, C., Solorio, T. (eds.) Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2-7, 2019, Volume 1 (Long and Short Papers). pp. 4171–4186. Association for Computational Linguistics (2019). https://doi.org/10.18653/v1/n19-1423, https://doi.org/10.18653/v1/n19-1423 7. Johnson, R.H., Blair, J.A.: Logical self-defense. International Debate Education Association (2006) 8. Potthast, M., Hagen, M., Stein, B., Graßegger, J., Michel, M., Tippmann, M., Welsch, C.: Chatnoir: a search engine for the clueweb09 corpus. In: Hersh, W.R., Callan, J., Maarek, Y., Sanderson, M. (eds.) The 35th International ACM SIGIR conference on research and development in Information Retrieval, SIGIR ’12, Portland, OR, USA, August 12-16, 2012. p. 1004. ACM (2012). https://doi.org/10.1145/2348283.2348429, https://doi.org/10.1145/2348283.2348429 9. Wachsmuth, H., Naderi, N., Hou, Y., Bilu, Y., Prabhakaran, V., Thijm, T.A., Hirst, G., Stein, B.: Computational argumentation quality assessment in natural language. In: Proceedings of the 15th Conference of the European Chapter of the Association for Computational Linguistics, EACL 2017, Valencia, Spain, April 3-7, 2017, Volume 1: Long Papers. pp. 176–187 (2017). https://doi.org/10.18653/v1/e17-1017, https://doi.org/10.18653/v1/e17-1017 10. Wachsmuth, H., Potthast, M., Khatib, K.A., Ajjour, Y., Puschmann, J., Qu, J., Dorsch, J., Morari, V., Bevendorff, J., Stein, B.: Building an argument search engine for the web. In: Habernal, I., Gurevych, I., Ashley, K.D., Cardie, C., Green, N., Litman, D.J., Petasis, G., Reed, C., Slonim, N., Walker, V.R. (eds.) Proceedings of the 4th Workshop on Argument Mining, ArgMining@EMNLP 2017, Copenhagen, Denmark, September 8, 2017. pp. 49–59. Association for Computational Linguistics (2017). https://doi.org/10.18653/v1/w17-5106, https://doi.org/10.18653/v1/w17-5106