=Paper=
{{Paper
|id=Vol-1178/CLEF2012wn-INEX-VivaldiEt2012
|storemode=property
|title=INEX Tweet Contextualization Track at CLEF 2012: Query Reformulation using Terminological Patterns and Automatic Summarization
|pdfUrl=https://ceur-ws.org/Vol-1178/CLEF2012wn-INEX-VivaldiEt2012.pdf
|volume=Vol-1178
}}
==INEX Tweet Contextualization Track at CLEF 2012: Query Reformulation using Terminological Patterns and Automatic Summarization==
INEX Tweet Contextualization Track at CLEF 2012: Query Reformulation using Terminological Patterns and Automatic Summarization Jorge Vivaldi and Iria da Cunha Universitat Pompeu Fabra Institut Universitari de Lingüı́stica Aplicada Barcelona {iria.dacunha,jorge.vivaldi}@upf.edu http://www.iula.upf.edu Abstract. The tweet contextualization INEX task at CLEF 2012 con- sists of the developing of a system that, given a tweet, can provide some context about the subject of the tweet, in order to help the reader to understand it. This context should take the form of a readable sum- mary, not exceeding 500 words, composed of passages from a provided Wikipedia corpus. Our general approach to get this objective is the fol- lowing: we perform some automatic reformulations of the initial tweets provided for the task (obtaining a list of terms related with the main topic of all them using terminological patterns). Then, using these re- formulated tweets, we obtain related documents with the search engine Indri. Finally, we use REG, an automatic extractive summarization sys- tem based on graphs, to summarize these documents and provide the summary associated to each tweet. Key words: INEX, CLEF, Tweets, Terms, Named Entities, Wikipedia, Auto- matic Summarization, REG. 1 Introduction The tweet contextualization INEX (Initiative for the Evaluation of XML Re- trieval) task at CLEF 2012 (Conference and Labs of the Evaluation Forum) consists of the developing of a system that, given a tweet, can provide some con- text about the subject of the tweet, in order to help the reader to understand it. This context should take the form of a readable summary, not exceeding 500 words, composed of passages from a provided Wikipedia corpus. Like in the Question-Answering (QA) of INEX 2011, the task to be performed by the partic- ipating groups is contextualizing tweets, that is answering questions of the form “what is this tweet about?” using a recent cleaned dump of the Wikipedia. The general process involves: tweet analysis, passage and/or XML elements retrieval and construction of the answer. Relevant passages would be segments contain- ing relevant information and also containing as little non-relevant information as possible (the result is specific to the question). The test data are about 1000 tweets in English collected by the organizers of the task from Twitter. They were selected among informative accounts (for example, @CNN, @TennisTweets, @PeopleMag, @science...), in order to avoid purely personal tweets that could not be contextualized. Information such as the user name, tags or URLs is provided. The document collection for all the participants, that is the corpus, has been rebuilt based on a dump of the English Wikipedia from November 2011. Resulting documents are made of a title, an abstract and sections with sub-titles. We consider that automatic extractive summarization systems could be useful in this QA task, taking into account that a summary can be defined as “a condensed version of a source document having a recognizable genre and a very specific purpose: to give the reader an exact and concise idea of the contents of the source” [1]. Summaries can be divided into “extracts”, if they contain the most important sentences extracted from the original text (ex. [2], [3], [4], [5], [6], [7]), and “abstracts”, if these sentences are re-written or paraphrased, generating a new text (ex. [8], [9], [10]). Most of the current automatic summarization systems are extractive. Our general approach is the following: we perform some automatic reformula- tions of the initial queries provided for the task (obtaining a list of terms related with the main topic of all the tweets using terminological patterns). Then, using these reformulated queries, we obtain related documents with the search engine Indri1 . Finally, we use REG ([11], [12]), an automatic extractive summarization system based on graphs, to summarize these documents and provide the final summary associated to each query. This approach is similar to the one used at QA@INEX track 2010 (see [13]) and 2011 (see [14]), since the same summarization system is employed. Never- theless, in our past participations, the system was semi-automatic, while in this work the system is totally automatic, from the reformulation of the queries us- ing terminological patterns, until the multi-document summarization of all the retrieved documents. The evaluation of the participant systems involves two aspects: informative- ness and readability. Informativeness evaluation is automatic, using the auto- matic evaluation system FRESA (FRamework for Evaluating Summaries Auto- matically) ([15], [16], [17]), and readability evaluation is carried out manually (evaluating syntactic incoherence, unsolved anaphora, redundancy, etc.). Following this introduction, the paper is organized as follows. In Section 2, the summarization system REG is shown. In Section 3, some information about terminology and terminological patterns is given. In Section 4, the methodol- 1 Indri is a search engine from the Lemur project, a cooperative work between the Uni- versity of Massachusetts and Carnegie Mellon University in order to build language modelling information retrieval tools: http://www.lemurproject.org/indri/ ogy is explained. In Section 5, experimental settings and results are presented. Finally, in Section 6, conclusions are exposed. 2 State-of-the-art and Resources 2.1 Term Extraction The notion of term that we have adopted in this work is based on the “Commu- nicative Theory of Terminology” [18]: a term is a lexical unit (single/multiple word) that activates a specialized meaning in a thematically restricted domain. Terms detection implies the distinction between domain-specific terms and gen- eral vocabulary. Its results are useful for any NLP task containing a domain specific component such as: ontology and (terminological) dictionary building, text indexing, automatic translation and summarization systems, among others. In spite of its large application field, its reliable and practical recognition still constitutes a bottleneck for many applications. As shown in [19], [20] and [21] among others, there are several methods to obtain the terms from a corpus. On the one hand, there are methods based on linguistic knowledge, like Ecode [22]. On the other hand, there are methods based on single statistical measures, such as ANA [23] or a combination of them, such as EXTERMINATOR [24]. Some tools combine both linguistic knowledge and statistically based methods, such as TermoStat [25], the algorithm shown in [26] or the bilingual extractors by [27] and [28]. However, none of these tools uses any kind of semantic knowledge. Notable exceptions are Metamap [29], Trucks [30] and YATE [31], among others. Also Wikipedia must be considered, since it is a very promising resource that is increasingly being used for both monolingual ([32], [33]) and multilingual term extraction [34]. Most of the tools, in particular those including an important linguistic compo- nent, takes into consideration the fact that terms usually follow a small number of POS patterns. In [35] it was shown that three patterns (noun, noun-adjective and noun-preposition-noun) cover more that 90% of the entries found in medical terminological dictionaries. Many of the above mentioned tools make some use of this fact. Nevertheless, some researchers like in [36] dynamically calculate the list of patterns found in terminological resources. 2.2 Named Entities Extraction Named Entity Recognition (NER) may be defined as the task to identify names referring to persons, organizations and locations in free text; later this task has been expanded to obtain other entities like dates and numeric expressions. This task was originally introduced as possible types of fillers in Information Extraction systems at the 6th Message Understanding Conference [37]. Although initially this task was limited to identify such expressions, later it has been expanded to their labeling with one entity type label (“person”, “organization”, etc.). Note that an entity (such as “Stanford”, the American university at the U.S.) can be referenced using several surface forms (e.g., “Stanford University” and “Stanford”) and a single surface form (e.g., “Stanford”) can refer to several entities (the university but also an American financer, several places in the UK or a financial group). See [38] for an interesting review. NER has proved to be a task useful for a number of NLP tasks as question an- swering, textual entailment and coreference resolution, among others. The recent interest in emerging areas like bioinformatics allows to expand this recognition task to proteins, drugs and chemical names. While early studies were mostly based on handcrafted rules, most recent ones use supervised machine learning as a way to automatically induce rule-based systems or sequence labeling algo- rithms starting from a collection of training examples. Often, corpus processing tools include some text handling facilities to perform simple NER detection for facilitating later processing. Some of them are based in language specific peculiarities such as initial upper case letters together with some heuristics for name entities placed at the beginning of the sentence. This is the case of the tool used for this experiment (see a description in [39]). 2.3 The REG System REG ([11], [12]) is an Enhanced Graph summarizer (REG) for extract sum- marization, using a graph approach. The strategy of this system has two main stages: a) to carry out an adequate representation of the document and b) to give a weight to each sentence of the document. In the first stage, the system makes a vectorial representation of the document. In the second stage, the system uses a greedy optimization algorithm. The summary generation is done with the con- catenation of the most relevant sentences (previously scored in the optimization stage). REG algorithm contains three modules. The first one carries out the vectorial transformation of the text with filtering, lemmatization/stemming and normal- ization processes. The second one applies the greedy algorithm and calculates the adjacency matrix. We obtain the score of the sentences directly from the algo- rithm. Therefore, sentences with a higher score are selected as the most relevant. Finally, the third module generates the summary, selecting and concatenating the relevant sentences. The first and second modules use CORTEX [6], a sys- tem that carries out an unsupervised extraction of the relevant sentences of a document using several numerical measures and a decision algorithm. 3 Methodology A main point in this research is to consider that named entities as well as words sequences that agree with the typical terminological patterns (see section 2.1) are representative of the tweets’ topic. To test this assertion, we design a method- ology to automatically retrieve all significant sequences from the tweets that satisfy the above mentioned criteria. The first step is to POS tag the tweets file. As a matter of fact, and in order to keep the process fully automatic, a minimal manipulation of the tweets file has been done. It includes only a minor modification to allow the text handling tool to keep the tweet id connected to the tweet itself. The next step, terminological patterns extraction, has been done using an already existent module of the YATE term extraction tool [31]. This information, together with the POS tagged tweet (to obtain proper nouns info) is used to build the query string for Indri. Some care has been taken to keep track of multiword sequences as indicated by the Indri query language specification (see examples below). In order to enrich the queries, we use a local installation of a Wikipedia dump2 to expand the terms with redirection information from such Wikipedia info. In this way, a query term like “Falklands” may be searched in the Wikipedia to find that it can be also referenced as “Falkland Islands”; therefore, the final query term is rewritten as: #syn(Falklands #1(Falkland Islands)) This strategy is also useful to find acronyms expansion as “USGS” and “United States Geological Survey” resulting in the following query: #syn("USGS" #1(United States Geological Survey)) Moreover, it allows to find words with different spellings as: #syn(#1(Christine de Pisan) #1(Christine de Pizan)) The resulting query has been delivered to Indri, using track organizer’s script, to obtain the Wikipedia pages relevant to every query. The following is an ex- ample of a full tweet: Increasingly, central banks, especially in emerging markets, have been the marginal buyers of gold http://t.co/9mftD5ju via WSJ. and its corresponding query string: #1(marginal buyers of gold),#1(emerging markets), #1(central banks),#syn("WSJ" #1(The Wall Street Journal)) The resulting set of Wikipedia pages has been split in several documents. Each document contains the pages relevant to the query. Such document is the input to the REG summarization system (see section 2.3), which builds a sum- mary with the significant passages. 2 This resource has been otained using [40]. 4 Experiments Settings and Results As mentioned in section 3, the process is fully automatic. No human intervention has taken place; therefore, errors and/or mistakes in the process may have a multiplicative effect. Most of such issues are exemplified as follows: 1. Tweet itself. The tweets file (including 1000 tweets) prepared by the organi- zation includes several errors like: mispelling, joined words, foreign language, etc. Consider the following examples: – 169657757870456833: “Lakers now 17-12 on the season & 12-2 at home. @paugasol 20pts 13rebs 4blks. Bynum 15pts 15rebs. @0goudelock 10pts, two 3 PTers.” – 169904294642978816: “@ranaoboy @Utcheychy @Jhpiego Thx for the #wiwchat RTs! Great conversation!” – 169655717538701312: “METTA. WORLD. PEACE.” – 170175722449670145: “http://t.co/amQ6IShA” – 170207412366745600: “RT @MexicanProblms: #41. When you’re eat- ing junk food y tu mom te dice que no comas "chucherias." #MexicanProblems”. Please note that, in some cases, it results in an empty query string or the resulting sentence is too short, causing POS tagging errors due to lack of context. 2. POS tagging. The output of most of the tools used for tagging (TreeTagger in this case) has some error rate. Unfortunately, errors mentioned above as well as extremely short sentences have a negative influence in the tagger performance. 3. Wikipedia expansion. It may happen that information added through Wikipe- dia expansion is not fully useful. This may be the case the only added infor- mation is the change of the case of some letters of the query term. 4. Indri query system. As shown in [41], this retrieval system has its own limits. 5. REG summarization system. The retrieval system issues a number of Wikipe- dia pages; therefore, it would be necessary to use a multidocument summa- rization system. As a matter of fact, REG is a single document summarizer, so some redundance may appear in the summaries. Some of the above issues may cause unusual results in the terminological patterns extraction tool. Therefore, in such cases, the pages retrieved by Indri may not correspond to the information available in Wikipedia about tweets’ topics. The evaluation of all the participant systems in the tweet contextualization INEX task at CLEF 2012 involves two aspects: informativeness and readability. On the one hand, as mentioned, to evaluate the informativeness the automatic FRESA package is used. This evaluation framework includes document-based summary evaluation measures based on probabilities distribution, specifically, the Kullback-Leibler (KL) divergence and the Jensen-Shannon (JS) divergence. As in the ROUGE package [42], FRESA supports different n-grams and skip n- grams probability distributions. FRESA environment has been used in the eval- uation of summaries produced in several European languages (English, French, Spanish and Catalan), and it integrates filtering and lemmatization in the treat- ment of summaries and documents. Table 1 includes the official results of the informativeness evaluation in the the tweet contextualization INEX task at CLEF 2012. This table presents the scores of the 33 participant runs. Table 1. Final results of informativeness in the tweet contextualization INEX task at CLEF 2012. Rank Run Uni Bi Skip 1 178 0.7734 0.8616 0.8623 2 152 0.7827 0.8713 0.8748 3 170 0.7901 0.8825 0.8848 4 194 0.7864 0.8868 0.8887 5 169 0.7959 0.8881 0.8904 6 168 0.7972 0.8917 0.8930 7 193 0.7909 0.8920 0.8938 8 185 0.8265 0.9129 0.9135 9 171 0.8380 0.9168 0.9187 10 186 0.8347 0.9210 0.9208 11 187 0.8360 0.9235 0.9237 12 154 0.8233 0.9254 0.9251 13 162 0.8236 0.9257 0.9254 14 155 0.8253 0.9280 0.9274 15 153 0.8266 0.9291 0.9290 16 196b 0.8484 0.9294 0.9324 17 196c 0.8513 0.9305 0.9332 18 196a 0.8502 0.9316 0.9345 19 164 0.8249 0.9365 0.9368 20 197 0.8565 0.9415 0.9441 21 163 0.8664 0.9628 0.9629 22 165 0.8818 0.9630 0.9634 23 150 0.9052 0.9871 0.9868 24 188 0.9541 0.9882 0.9888 25 176 0.8684 0.9879 0.9903 26 149 0.9059 0.9916 0.9916 27 156 0.9366 0.9913 0.9916 28 157 0.9715 0.9931 0.9937 29 191 0.9590 0.9947 0.9947 30 192 0.9590 0.9947 0.9947 31 161 0.9757 0.9949 0.9950 32 177 0.9541 0.9981 0.9984 33 151 0.9223 0.9985 0.9988 As shown in Table 1, our run (165) obtains the position 22 in the rank. Exactly, it obtains 0.8818 using unigrams, 0.9630 using bigrams and 0.9634 using skip bigrams. The best run in the ranking (178) obtains 0.7734, 0.8616 and 0.8623, respectively. On the other hand, readability is evaluated manually. Evaluators are asked to evaluate several aspects related to syntactic incoherence, unsolved anaphora, redundancy, etc. The specific orders given to evaluators are: – Syntax S: “Tick the box is the passage contains a syntactic problem (bad segmentation for example)”. – Anaphora A: “Tick the box if the passage contains an unsolved anaphora”. – Redundancy R: “Tick the box if the passage contains a redundant informa- tion, i.e. an information that have already been given in a previous passage”. – Trash T: “Tick the box if the passage does not make any sense in its context (i.e. after reading the previous passages). These passages must then be con- sidered as trashed, and readability of following passages must be assessed as if these passages were not present”. The score is the average normalized number of words in valid passages, and participants are ranked according to this score. Summary word numbers are normalized to 500 words each. Table 2 includes the final results of readability evaluation in the tweet con- textualization INEX task at CLEF 2012. Estimated average scores are available for: – Relevance: proportion of text that makes sense in context. – Syntax: proportion of text without syntax problems. – Structure: proportion of text without broken anaphora and avoiding redun- dancy. These measures were estimated on the same pool of tweets as for previously released informativeness evaluation by organizers. Runs that failed to provide at least 6 consistent summaries in this pool have been kept apart because the estimates were too uncertain for inclusion in the official results. Because of this reason, in Table 2 only 27 runs are shown. As shown in Table 2, our run (165) obtains the position 7 in the rank. Exactly, it obtains 0.5936 using unigrams, 0.6049 using bigrams and 0.5442 using skip bigrams. The best run in the ranking (185) obtains 0.7728, 0.7452 and 0.6446, respectively. These results show that the performance of our system is not so good regard- ing informativeness, but it is much better regarding readability. This difference between informativeness and readability is also shown by other systems (see for example the best runs in both categories, 178 and 185). In our case, we consider that the mentioned mistakes in the tweets and the fact that the terminology extraction is totally automatic can cause that the pages retrieved by Indri are not as relevant as expected. Nevertheless, using an automatic summarization system, we can guarantee that the quality of readability is acceptable. Table 2. Final results of readability in the tweet contextualization INEX task at CLEF 2012. Run Relevance Syntax Structure 185 0.7728 0.7452 0.6446 171 0.631 0.606 0.6076 168 0.6927 0.6723 0.5721 194 0.6975 0.6342 0.5703 186 0.7008 0.6676 0.5636 170 0.676 0.6529 0.5611 165 0.5936 0.6049 0.5442 152 0.5966 0.5793 0.5433 155 0.6968 0.6161 0.5315 178 0.6336 0.6087 0.5289 169 0.5369 0.5208 0.5181 193 0.6208 0.6115 0.5145 163 0.5597 0.555 0.4983 187 0.6093 0.5252 0.4847 154 0.5352 0.5305 0.4748 196b 0.4964 0.4705 0.4204 153 0.4984 0.4576 0.3784 164 0.4759 0.4317 0.3772 162 0.4582 0.4335 0.3726 197 0.5487 0.4264 0.3477 196c 0.449 0.4203 0.3441 196a 0.4911 0.3813 0.3134 176 0.2832 0.2623 0.2388 156 0.2933 0.2716 0.2278 188 0.1542 0.1542 0.1502 157 0.1017 0.1045 0.1045 161 0.0867 0.0723 0.0584 5 Conclusions and Future Work In this paper, our strategy and results for the tweet contextualization INEX task at CLEF 2012 are presented. The task consists of the developing of a system that, given a tweet, can provide some context about the subject of the tweet, in order to help the reader to understand it. This context should take the form of a readable summary, not exceeding 500 words, composed of passages from a provided Wikipedia corpus. The test data are about 1000 tweets in English collected by the organizers of the task from Twitter. Our system performs some automatic reformulations of the initial tweets pro- vided for the task (obtaining a list of terms related with their main topic using terminological patterns). Then, using these reformulated tweets, we obtain re- lated documents with the search engine Indri. Finally, we use REG to summarize these documents and provide the final summary associated to each tweet. The results show that, comparing to the other participants, the performance of our system is not so good regarding informativeness (probably due to mistakes in the tweets and problems in the terminology extraction process), but it is much better regarding readability (probably due to the fact of using a summarization system). In the future we plan to follow several parallel lines: i) to improve term selection and its expansion to refine the queries and therefore to improve the pertinence of the Wikipedia pages retrieved by Indri; ii) to further investigate the actual pertinence of the Wikipedia retrieved pages to the query; and iii) to check the actual weight of summarization process in the full task by testing other summarization systems. References 1. Saggion, H.; Lapalme, G. (2002). Generating Indicative-Informative Summaries with SumUM. Computational Linguistics 28(4). 497-526. 2. Edmunson, H. P. (1969). New Methods in Automatic Extraction. Journal of the Association for Computing Machinery 16. 264-285. 3. Nanba, H.; Okumura, M. (2000). Producing More Readable Extracts by Revising Them. In Proceedings of the 18th Int. Conference on Computational Linguistics (COLING-2000). Saarbrucken. 1071-1075. 4. Gaizauskas, R.; Herring, P.; Oakes, M.; Beaulieu, M.; Willett, P.; Fowkes, H.; Jonsson, A. (2001). Intelligent access to text: Integrating information extraction technology into text browsers. In Proceedings of the Human Language Technology Conference. San Diego. 189-193. 5. Lal, P.; Reger, S. (2002). Extract-based Summarization with Simplication. In Pro- ceedings of the 2nd Document Understanding Conference at the 40th Meeting of the Association for Computational Linguistics. 90-96. 6. Torres-Moreno, J-M.; Velázquez-Morales, P.; Meunier, J. G. (2002). Condensés de textes par des méthodes numériques. In Proceedings of the 6th Int. Conference on the Statistical Analysis of Textual Data (JADT). St. Malo. 723-734. 7. da Cunha, I.; Fernández, S.; Velázquez, P.; Vivaldi, J.; SanJuan, E.; Torres-Moreno, J-M. (2007). A new hybrid summarizer based on Vector Space Model, Statistical Physics and Linguistics. Lecture Notes in Computer Science 4827. 872-882. 8. Ono, K.; Sumita, K.; Miike, S. (1994). Abstract generation based on rhetorical structure extraction. In Proceedings of the Int. Conference on Computational Lin- guistics. Kyoto. 344-348. 9. Paice, C. D. (1990). Constructing literature abstracts by computer: Techniques and prospects. Information Processing and Management 26. 171-186. 10. Radev, D. (1999). Language Reuse and Regeneration: Generating Natural Lan- guage Summaries from Multiple On-Line Sources. PhD Thesis. New York, Columbia University. 11. Torres-Moreno, J-M.; Ramı́rez, J. (2010). REG : un algorithme glouton appliqué au résumé automatique de texte. Proceedings of the 10th Int. Conference on the Statistical Analysis of Textual. Roma, Italia. 12. Torres-Moreno, J-M.; Ramı́rez, J.; da Cunha, I. (2010). Un resumeur a base de graphes, indépendant de la langue. In Proceedings of the Int. Workshop African HLT 2010. Djibouti. 13. Vivaldi, J.; da Cunha, I.; Ramı́rez, J. (2011). The REG summarization system with question reformulation at QA@INEX track 2010. In Geva, S. et al. (eds.). INEX 2010, Lecture Notes in Computer Science 6932. 295-302. Berlı́n: Springer. 14. Vivaldi, J.; da Cunha, I. (2012). QA@INEX Track 2011: Question Expansion and Reformulation Using the REG Summarization System. Lecture Notes in Computer Science (LNCS) 7424. 257-268. Berlin: Springer. 15. Saggion, H.; Torres-Moreno, J-M.; da Cunha, I.; SanJuan, E.; Velázquez-Morales, P.; SanJuan, E. (2010). Multilingual Summarization Evaluation without Human Models. In Proceedings of the 23rd Int. Conference on Computational Linguistics (COLING 2010). Pekin. 16. Torres-Moreno, J-M.; Saggion, H.; da Cunha, I.; SanJuan, E.; Velázquez-Morales, P. (2010). Summary Evaluation With and Without References. Polibitis: Research journal on Computer science and computer engineering with applications 42. 17. Torres-Moreno, J-M.; Saggion, H.; da Cunha, I.; Velázquez-Morales, P.; SanJuan, E. (2010). Ealuation automatique de résumés avec et sans référence. In Proceed- ings of the 17e Conférence sur le Traitement Automatique des Langues Naturelles (TALN). Montreal: Univ. de Montréal et Ecole Polytechnique de Montréal. 18. Cabré, M. T. (1999). La terminologı́a. Representación y comunicación. Barcelona: IULA. 19. Cabré, M. T.; Estopà, R.; Vivaldi, J. (2001). Automatic term detection. A review of current systems. Recent Advances in Computational Terminology 2. 53-87. 20. Pazienza, M. T.; Pennacchiotti, M.; Zanzotto, F.M. (2005). Terminology Extrac- tion: An Analysis of Linguistic and Statistical Approaches. Studies in Fuzziness and Soft Computing 185. 255-279. 21. Ahrenberg, L. (2009). Term Extraction: A Review. (Unpublished draft). 22. Alarcón, R.; Sierra, G.; Bach, C. (2008). ECODE: A Pattern Based Approach for Definitional Knowledge Extraction. In Proceedings of the XIII EURALEX Int. Congress. Barcelona: IULA, UPF, Documenta Universitaria. 923-928. 23. Enguehard, C.; Pantera, L. (1994). Automatic Natural Acquisition of a Terminol- ogy. Journal of Quantitative Linguistics 2(1). 27-32. 24. Patry, A.; Langlais, P. (2005). Corpus-based terminology extraction. In Proceedings of 7th Int. Conference on Terminology and Knowledge Engineering. Copenhagen. 25. Drouin, P. (2002). Acquisition automatique des termes: l’utilisation des pivots lex- icaux spécialisés. Ph.D. Thesis. Montreal (Canada): Université de Montréal. 26. Frantzi, K. T.; Ananiadou, S.; Tsujii, J. (2009). Erdmann, M.; Nakayama, K.; Hara, T.; Nishio, S. (2009). The C-value/NC-value Method of Automatic Recognition for Multi-word Terms. Lecture Notes in Computer Science 1513. 585-604 27. Vintar, S. (2010). Bilingual term recognition revisited: The bag-of-equivalents term alignment approach and its evaluation. Terminology 16(2). 141-158. 28. Gómez Guinovart, X. (2012). A Hybrid Corpus-Based Approach to Bilingual Ter- minology Extraction. In I. Moskowich and B. Crespo (eds.). Encoding the Past, Decoding The Future: Corpora in the 21st Century. Cambridge Scholar Publish- ing: Newcastle upon Tyne. 147-175. 29. Aronson, A.; Lang, F. (2010). An overview of MetaMap: historical perspective and recent advances. Journal of the American Medical Informatics Association 17(3). 229-236. 30. Maynard, D. (1999). Term Recognition Using Combined Knowledge Sources. Ph.D. Thesis. Manchester Metropolitan University. Manchester (UK). 31. Vivaldi, J. (2001). Extracción de candidatos a término mediante combinación de estrategias heterogéneas. Ph.D. thesis. Universitat Politècnica de Catalunya. Barcelona (Spain). 32. Vivaldi, J.; Rodrı́guez, H. (2010). Using Wikipedia for term extraction in the biomedical domain: first experiences. Procesamiento del Lenguaje Natural 45. 251- 254. 33. Cabrera-Diego, L.; Sierra, G.; Vivaldi, J.; Pozzi, M. (2011). Using Wikipedia to Validate Term Candidates for the Mexican Basic Scientific Vocabulary. In Pro- ceedings of LaRC 2011: First Int. Conference on Terminology, Languages, and Content Resources. Seoul. 76-85. 34. Erdmann, M.; Nakayama, K.; Hara, T.; Nishio, S. (2009). Improving the Extrac- tion of Bilingual Terminology from Wikipedia. ACM Transactions on Multimedia Computing, Communications and Applications 5(4). 31.1-31.16. 35. Estopà, R. (1999). Extracting Lexical Semantic Knowledge from Wikipedia and Wiktionary. Ph.D. Thesis. Pompeu Fabra University. Barcelona (Spain). 36. Nazar, R.; Cabré, M. T. (2012). Supervised Learning Algorithms Applied to Ter- minoloy Extraction. In 10th Terminology and Knowledge Engineering Conference. 37. Grishman, R.; Sundheim, B. (1996). Message Understanding Conference - 6: A Brief History. In Proceedings of the 16th Int. Conference on Computational Lin- guistics. 466-471. 38. Nadeau, D.; Sekine, S. (2007). A survey of named entity recognition and classifi- cation. Journal of Linguisticae Investigationes 30(1). 3-26. 39. Martı́nez, H.; Vivaldi, J.; Villegas, M. (2010). Text handling as a Web Service for the IULA processing pipeline. In Proceedings of the 7th conference on International Language Resources and Evaluation (LREC’10). 22-29. 40. Zesch, T.; Müller, C., Gurevych, I. (2008). Extracting Lexical Semantic Knowledge from Wikipedia and Wiktionary. In 6th LREC Conference Proceedings. 1646-1652. 41. Strohman, Trevor; Metzler, Donald; Turtle, Howard; Croft, Bruce (2005). Indri: A language-model based search engine for complex queries. University of Mas- sachusetts Amherst. CIIR Technical Report IR-407. 42. Lin, C-Y. (2004). ROUGE: A Package for Automatic Evaluation of Summaries. In Proceedings of Text Summarization Branches Out: ACL-04 Workshop. 74-81.