=Paper=
{{Paper
|id=Vol-1175/CLEF2009wn-QACLEF-BharadwajEt2009
|storemode=property
|title=A Naïve Approach for Monolingual Question Answering
|pdfUrl=https://ceur-ws.org/Vol-1175/CLEF2009wn-QACLEF-BharadwajEt2009.pdf
|volume=Vol-1175
|dblpUrl=https://dblp.org/rec/conf/clef/BharadwajV09
}}
==A Naïve Approach for Monolingual Question Answering==
1 A Naïve Approach for Monolingual Question Answering Rohit Bharadwaj, Surya Ganesh, Vasudeva Varma, {bharadwaj, suryag}@research.iiit.ac.in, vv@iiit.ac.in LTRC, IIIT-Hyderabad track with both source and destination languages as English. We indexed the data using Lucene, an open source search library. Lucene implements okapi BM25 retrieval model [6]. Abstract— This paper talks about the system which we have Using this search library we have built passage index, that is, submitted for the ResPubliQA task. We participated in building each passage in a document is considered as a retrieval unit. the QA system for en-en part. We followed a different method for each question type. In this paper we outline the methods which we adapted and the results which we obtained. 3 OUR APPROACH Our QA system incorporates pipeline architecture as shown in figure 1. It consists of three core components: 1) Question 1 INTRODUCTION analysis, 2) Passage retrieval and 3) Passage selection. The implementation details of all the three components are T HE main focus of QA is to gain the knowledge of the user’s question and retrieve the sentences that are described below. close to the answer. The ResPubliQA task expects the system to understand the question and retrieve the corresponding Question passage in the text which contains the answer. The architecture of our QA system is 1) Question Analysis 2) Passage retrieval and 3) Passage selection. Question analysis Question Analysis involves the classification of the question into pre-defined question types, extraction of query words and determining the answer type. Passage retrieval searches for passages in the Passage Retrieval document collection which are likely to contain the answer. Passage selection ranks the list of candidate answers to determine the final answer. First, we introduce the task, and then we describe the pre-processing we carried over the data Passage Selection in section 2. In section 3 we describe the approach which we have followed for answering the questions. Section 4 describes the results we obtained while section 5 presents the Answer analysis and conclusion. Figure 1: Pipeline Architecture of our QA system 2 PREPROCESSING THE DATA Question Analysis In question analysis, we classified the given 95 questions into In ResPubliQA task, we are provided with the data that is one of the pre-defined classes. The pre-defined question types delimited into passages and we are expected to return the or classes are Factoid, Definitive, Reason, Procedure, passage that contains the answer. We are provided with both purpose. The classification is semi-automatic. As the question the question language and the target language in which the classes are fixed, by observation we identified patterns for answers are to be present. The task is mainly directed towards each of the classes. The patterns for Factoid and Definitive cross language question answering. We participated in the are inter-related and hence we classified the questions under these categories into a single class FactDef which was later 2 sub-divided into factoid and definitive classes. The observed an accuracy of 78.2% under fine grained patterns are shown in Table 1. classification. As the classification accuracy is higher for coarse grained classification and also because of the limitations of many NER systems to Question class Words that Words that Number of recognize fine grained named entity types in (Answer type) must be should not be questions passages, only coarse answer type is used to identify present in the present in the passages with answer candidates. question (case question (case 2. Density: Tellex et al. [5] showed that density based folded) folded) measures work well for passage retrieval in QA. So, FactDef what, how, aim, goal, 46 the passages resulting from the above step are then defin, who, objective, re-ranked based on the density of the question where, name reason, keywords in them. Density is defined as the average procedure, distance between the answer and question keywords purpose in a passage. There are several ways to compute Reason why, reason - 33 density. We adopt a simple formula as described in Procedure Procedure - 10 [4] to compute density of query terms in a passage. Purpose aim, objective, - 6 goal, purpose Finally, among the re-ranked passages, the top ranked passage is produced as the answer given a question. Table 1: Generic patterns in questions from different classes Definitive Out of the 46 questions in FactDef class, 27 questions are from factoid class and remaining 19 are from definitive class. We used answer patterns for definitive questions and used them for passage selection. The question focus or Qword are The methods followed for each of the question class which extracted by removing the stop words (a pre-compiled list) includes both passage retrieval and passage selection from the question. methodologies are described below. The main answer patterns for definitive questions as given in Factoid [1] are (where A is the Qword and X is the expected answer) 1) To answer a factoid question, first, we retrieve a set of relevant passages. So, a keyword query constructed by 2) stripping of all the stop words and interrogative words (when, where, which etc.) in the question. This keyword query isgiven to Lucene to retrieve a ranked set of relevant passages. From this set, one of the passages is given as the answer to a 3) question. Our approach for selecting an answer containing 4) passage is a two step process as described below. < X; [comma]; [also] called; A [comma]> 1. Answer type: Using the answer type of a question, we identify a set of passages which contain answer candidates. To obtain the answer type of a question, we have implemented a question classifier using 5) support vector machines (SVM) [3]. The classifier 6) was trained on UIUC [2] dataset which consists of 5,500 questions for training and 500 questions for As our system does not need to extract the answer but to testing. Every question in the dataset was labeled retrieve the passage, we modified the patterns and extended into a coarse grained and a fine grained category them by adding few more patterns like “Qword + means/ from a total number of 6 and 50 categories mean/ has/”. So effectively the queries used to search the respectively. We have used the bag-of-words feature index are the modified queries which are formed by adding to predict the category, that is, the answer type of a the answer patterns. We also queried the index by adding question. The classifier showed an accuracy of various versions of the modified query like "Qword means", 86.8%, when tested on 500 questions from the UIUC Qword + means, “Qword, called” etc. This resulted in various dataset under the coarse grained classification. And, results for each modified query. For identifying the correct 3 answer, we performed various experiments like giving boost Total questions 95 to results of a particular query, giving weight to each query and calculating the final weight of each result, performing Question answered correctly 54 various intersection and union operations for finding the final Incorrectly answered 37 result on the development dataset. From these experiments, questions the one that gave most correct results was “prioritizing the Questions unanswered 4 search patterns and preserving that order while searching”. The optimal priority order of search patterns is shown below. Table 2: Results for monolingual English ResPubliQA task 1. "Qword means" 2. “Qword mean" 5 CONCLUSION 3. “Qword is/are" We described our participation in ResPubliQA task. We have developed a monolingual English QA system. Our system 4. “Qword, called” does not rely on any external knowledge resources or any complex information retrieval, information extraction and 5. “called Qword” natural language processing techniques. Instead, it uses an 6. “Qword was” effective combination of naive techniques from the above areas to achieve a decent performance. Our system analyses 7. Qword passages from passage retrieval output to identify the correct answer in the case of factoid and definition questions, whereas, the top ranked passage was produced as an answer So if we achieve result for the first query, then it is given as for the remaining questions in the test set. The analysis the final answer. If there are no results then we proceed to the method used for selecting the answer differs for factoid and next query. Finally if we have no results for any of the definition questions. patterns, then the system doesn't answer the question. All the remaining questions, that is, questions from Reason, 6 References Procedure and Purpose types are answered naively by giving a top ranked passage, with a minimum length of N words, from [1] Soubbotin Soubbotin Insightsoft-M. Patterns of the passage retrieval component as the answer. In our Potential Answer Expressions as Clues to the Right experiments on the development data, we have observed Answers. In Proceedings of the Tenth Text REtrieval better results for N=25. So, we used the same value for Conference (TREC), 2001. finding the answer to the questions in test data. [2] Xin Li Dan, Xin Li, and Dan Roth. 2002. Learning question classifiers. pages 556–562. 4 RESULTS [3] Zhang, Dell and Lee, Wee Sun. Question classification using support vector machines. In proceedings of the 26th The test dataset for ResPubliQA task consists of a subset of annual international ACM SIGIR conference on Research the JRC-ACQUIS Multilingual Parallel Corpus4, and 500 and development in information retrieval, 2003. questions distributed over factoid, definitive, procedure, reason and purpose classes. JRC-ACQUIS is a freely available [4] G. G. Lee, J. Seo, S. Lee, H. Jung, B.-H. Cho, C. Lee, B.-K. parallel corpus of European Union legal documents. It Kwak, J. Cha, D. Kim, J. An, H. Kim, and K. Kim. SiteQ: Engineering high performance QA system using lexico- comprises selected texts written between 1950 and 2006 with semantic pattern matching and shallow NLP. In parallel translation in 22 European languages. We used Proceedings of the Tenth Text REtrieval Conference English documents in the corpus. Out of the 500 questions in (TREC 2001), 2001. all languages, we have answered 95 questions which are categorized under monolingual English QA task. The results [5] Stefanie Tellex and Boris Katz and Jimmy Lin and Aaron of our system as provided by the CLEF are shown in table 2. Fernandes and Gregory Marton. Quantitative Evaluation of Passage Retrieval Algorithms for Question Answering. In Proceedings of the 26th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR), 2003. [6] S. E. Robertson, S. Walker, M. Hancock-Beaulieu, M. Gatford, and A. Payne. Okapi at TREC-4. In Proceedings of the 4th Text REtrieval Conference (TREC-4), 1995.