=Paper=
{{Paper
|id=Vol-1171/CLEF2005wn-WebCLEF-PintoEt2005
|storemode=property
|title=TPIRS: A System for Document Indexing Reduction on WebCLEF
|pdfUrl=https://ceur-ws.org/Vol-1171/CLEF2005wn-WebCLEF-PintoEt2005.pdf
|volume=Vol-1171
|dblpUrl=https://dblp.org/rec/conf/clef/PintoJRS05a
}}
==TPIRS: A System for Document Indexing Reduction on WebCLEF==
TPIRS: A System for Document Indexing
Reduction on WebCLEF∗
David Pinto & Héctor Jiménez-Salazar
Faculty of Computer Science, BUAP
{dpinto, hjimenez}@cs.buap.mx
Paolo Rosso & Emilio Sanchis
Department of Information Systems and Computation, UPV
{prosso, esanchis}@dsic.upv.es
Abstract
In this paper we present the results of BUAP/UPV universities in WebCLEF, a par-
ticular task of CLEF 2005. Particularly, we evaluate our information retrieval system
in the bilingual English to Spanish track. Our system uses a term reduction process
based on the Transition Point technique. Our results show that it is possible to re-
duce the number of terms to index, thereby improving the performance of our system.
We evaluate different percentages of reduction over a subset of EuroGOV, in order
to determine the best one. We observed that after reducing the 82.55% of the cor-
pus, a Mean Reciprocal Rank of 0.0844 was obtained, compared with 0.0465 of such
evaluation with full documents.
Categories and Subject Descriptors
H.3 [Information Storage and Retrieval]: H.3.1 Content Analysis and Indexing; H.3.3 Infor-
mation Search and Retrieval; H.3.4 Systems and Software; H.3.7 Digital Libraries
General Terms
Measurement, Performance, Experimentation
Keywords
Cross-Lingual Information Retrieval, Terms Reduction, Transition Point
1 Introduction
High volume of information in Internet leds to developed novel techniques for managing of data,
specially when we deal with information in multiple languages. There are sufficient example
scenarios in which users may be interested in information which is in a language other than their
own native language. A common language scenario is where a user has some comprehension ability
for a given language but s/he is not sufficiently proficient to confidently specify a search request
in that language. Thus, a search system that can deal with this problem should be of a high
benefit. The World Wide Web (WWW) is a natural setting for cross-lingual information retrieval;
∗ This work was partially supported by BUAP-VIEP 3/G/ING/05, R2D2 (CICYTTIC2003-07158-C04-03), ICT
EU-India (ALA/95/23/2003/077-054) research projects
the European Union is a typical example of a multilingual scenario, where multiple users have to
deal with information published in at least 20 languages.
In order to reinforce research in this area, CLEF (Cross-Language Evaluation Forum) has been
compiling a set of multi-lingual corpora and promoting the evaluation of multiple multi-lingual
information retrieval systems for diverse kinds of data [4]. A particular track for the evaluation of
such systems that deal with information on the web has been set up this year as a part of CLEF.
This forum was named WebCLEF, and the best description of this particular task can be seen in
[10]. In WebCLEF, three subtasks were defined within this year: mixed monolingual, multilingual,
and bilingual English to Spanish.
This paper reports results on the evaluation of a Cross-Language Information Retrieval System
(CLIRS) for the bilingual English to Spanish subtask of WebCLEF 2005. A document indexing
reduction is proposed, in order to improve precision of CLIRS and to diminish the storing space
on such systems. Our proposal is based on the use of the Transition Point (TP) technique, which
is somehow a method that obtains important terms from a document. We evaluate different
percentages of TP over a subset of EuroGOV corpus [9], and we observed that it is possible to
improve precision results reducing the number of terms for a given corpus.
The next section describes our information retrieval system in detail. Section 3 briefly intro-
duces the corpus used in our experiments, and the results obtained after evaluation. Finally, a
discussion of our experiments is presented.
2 Description of TPIRS
We used a boolean model with Jaccard similarity formula for our CLIRS. Our goal was to determine
the behaviour of document indexing reduction in an information retrieval environment. In order
to reduce the terms from every document treated, we applied a technique named Transition Point,
which is described as follows.
2.1 Transition Point Technique
The transition point is a frequency value that splits the vocabulary of a document in two sets of
terms (low and high frequency). This technique is based on Zipf Law of Word Ocurrences [14]
and refined from studies of Booth [1] and, recently, Urbizagástegui [13]. These studies are meant
to demonstrate that terms of medium frequency are closely related to the conceptual content of
the document. Thus, it is possible to form the hypothesis that terms closer to TP can be used as
indexes of a document. A typical formula used to obtain this value is given in equation 1.
√
8 ∗ I1 + 1 − 1
TP = , (1)
2
where I1 represents the number of words with frequency equal to 1 [8] [13].
Alternatively, TP can be localized by identifying the lowest frequency (from the highest fre-
quencies) that it is not repeated; this characteristic comes from properties of Booth’s law of low
frequency words [1].
Let us consider a frequency-sorted vocabulary of a document; i.e., VT P = [(t1 , f1 ), ..., (tn , fn )],
with fi ≥ fi−1 , then T P = fi−1 , iif fi = fi+1 . The most important words are those that obtain
the closest frequency values to TP, i.e.,
T PSET = {ti |(ti , fi ) ∈ VT P , U1 ≤ fi ≤ U2 }, (2)
where U1 is a lower threshold obtained by a given neighbourhood percentage of TP (NTP),
thus, U1 = (1 − N T P ) ∗ T P . U2 is the upper threshold and it is calculated in a similar way
(U2 = (1 + N T P ) ∗ T P ).
We have used TP technique in diverse areas of natural language processing (NLP), like: clus-
tering of short texts [5], categorization of texts [6], keyphrases extraction [7] [12], summarization
[2], and weighting models for information retrieval systems [3]. Thus, we believe that there exist
enough evidence to utilize this technique as a terms reduction process.
2.2 Information Retrieval Model
Our information retrieval is based on the Boolean Model, and, in order to rank the documents
retrieved, we used the Jaccard similarity function, applied to the query and every document of the
corpus used. Previously, each document was preprocessed and its index terms were selected (the
preprocessing phase is described in section 3.1). For this purpose, several values of a neighbourhood
of TP were used as thresholds, as equation 2 indicates.
3 Evaluation
3.1 Corpus
We used a subset of EurGov corpus for our evaluation. This subset was composed by a set of
Spanish Internet pages, originally obtained from European government-related sites. We named
this corpus BiEnEs.
In order to construct this corpus, for every page compiled in the EuroGOV corpus, we determine
its language by using TexCat [11], a guesser language program widely used. We construct our
evaluation corpus with those documents identified as Spanish language.
The preprocessing of the BiEnEs corpus consisted of elimination of punctuation symbols, Span-
ish stopwords, numbers, html tags, script codes and style cascade sheets codes.
For the evaluation of BiEnEs, a set of 134 queries was composed and refined, in order to provide
gramatically correct “English” queries. Queries and assessments were created by the participants
in the WebCLEF track, and the particular case of the queries were later reviewed and in some
cases corrected in their English translation by the NLP Group at UNED. Queries were distributed
in the following way: 67 homepages and 67 named page findings.
We applied a preprocessing phase to this set of queries. First, we used an online translation
system 1 in order to translate every query from English to Spanish. After that, an elimination of
punctuation symbols, spanish stopwords and numbers was done.
We did not apply a rigorous method of translation, due to the fact that our main goal in our
first participation on WebCLEF was to determine the quality of terms reduction in our CLIRS.
3.2 Indexing reduction
In order to determine the behaviour of document indexing reduction on CLIRS, we submit to the
contest, a set of five runs, which are described as follows.
First Run: This run used “Full documents” as evaluation corpus, and conformed the baseline
for our experiments. We named it the “Full” evaluation.
Second Run: This run used an evaluation corpus composed of the reduction of every document,
using the TP technique with a neighbourhood of 10% around TP. We named it the “TP10”
evaluation.
Third Run: This run used an evaluation corpus composed of the reduction of every document,
using the TP technique with a neighbourhood of 20% around TP. We named it the “TP20”
evaluation.
Fourth Run: This run used an evaluation corpus composed of the reduction of every document,
using the TP technique with a neighbourhood of 40% around TP. We named it the “TP40”
evaluation.
1 http://www.freetranslation.com
Fifth Run: This run used an evaluation corpus composed of the reduction of every document,
using the TP technique with a neighbourhood of 60% around TP. We named it the “TP60”
evaluaton.
Table 1 shows the size of every evaluation corpus used, as well as the percentage of reduction
obtained for each one. As can be seen, the TP technique obtained a big percentage of reduction
(between 75 and 89%), which also implies a reduction in time for indexing process, in a CLIRS.
Corpus Size (Kb) % of Reduction
Full 117,345 0%
TP10 12,616 89.25%
TP20 19,660 83.25%
TP40 20,477 82.55%
TP60 28,903 75.37%
Table 1: Evaluation corpora
3.3 Results
Table 2 shows the results for every run submitted. First column indicates the name of each run.
Last column shows the Mean Reciprocal Rank (MRR) obtained for each run. Additionally, the
average success at different number of documents retrieved is shown, by instance, second column
indicates the average success of the CLIRS at the first answer. The “TP20” approach, obtained a
total of 49 answers, and therefore, it does not has average success at 50.
As can be seen, an important improvement was done by using an evaluation corpus obtained
with a neighbourhood of 40% of TP. We were hoping to obtain comparable results with the “Full”
run, but as can be seen, the “TP40” approach duplicated “Full” MRR.
Average Success at
Corpus 1 5 10 20 50 Mean Reciprocal Rank
Full 0.0224 0.0672 0.1119 0.1418 0.1866 0.0465
TP10 0.0224 0.0373 0.0672 0.0821 0.1119 0.0331
TP20 0.0299 0.0448 0.0672 0.1045 – 0.0446
TP40 0.0597 0.0970 0.1119 0.1418 0.2164 0.0844
TP60 0.0522 0.1045 0.1269 0.1642 0.2090 0.0771
Table 2: Evaluation results
4 Discussion
We proposed an index reduction method for a cross-lingual information retrieval system. Our
proposal is based on the transition point technique.
After submitting five runs on the bilingual English to Spanish subtrack from WebCLEF, we
observed that it is possible to reduce terms in the documents that conform the corpus of a CLIRS,
not only by reducing the time needed for indexing but also by improving the precision of the
results obtained by CLIRS.
Our method is linear in computational time, and therefore it can be used in practical tasks.
Until now, results obtained in terms of MRR are very low, but findings show that by applying better
techniques of English to Spanish translation of queries, results can be dramatically improved.
We were concerned with the impact of indexing reduction on CLIRS, and in the future we
hope to improve other components of our CLIRS, for instance, the use of vector space model, in
order to improve the MRR.
The TP technique has shown an effective use on diverse areas of NLP, and its best features
for NLP, are mainly two: a high content of semantic information and the sparseness that can
be obtained on vectors for document representation on models based on the vector space model.
On the other hand, its language independence allows to use this technique in CLIRS, that is the
matter of WebCLEF.
References
[1] A. Booth: A Law of Ocurrences for Words of Low Frequency, Information and control, 1967.
[2] C. Bueno, D. Pinto, H. Jimenez, El párrafo virtual en la generación de extractos, Research
on Computing Science Journal, ISSN 1665-9899, 2005.
[3] R. Cabrera, D. Pinto, H. Jimenez, D. Vilariño, Una nueva ponderación para el modelo de
espacio vectorial de recuperación de información, Research on Computing Science Journal,
ISSN 1665-9899, 2005.
[4] CLEF 2005: Cross-Language Evaluation Forum, http://www.clef-campaign.org/, 2005.
[5] H. Jimenez, D. Pinto, P. Rosso, Selección de Términos No Supervisada para Agrupamiento
de Resúmenes, In proceedings of Workshop on Human Language, ENC05, 2005.
[6] E. Moyotl, H. Jimenez, An Analysis on Frequency of Terms for Text Categorization, Pro-
ceedings of XX Conference of Spanish Natural Language Processing Society (SEPLN-04),
2004.
[7] D. Pinto, F. Pérez:Una Técnica para la Identificación de Términos MultipalabrIn p, Proceed-
ings of 2nd. National Conference on Computer Science, Mexico, 2004.
[8] B. Reyes-Aguirre, E. Moyotl-Hernández & H. Jiménez-Salazar.: Reducción de Términos In-
dice Usando el Punto de Transición, In proceedings of Facultad de Ciencias de Computación
XX Anniversary Conferences, BUAP, 2003.
[9] B. Sigurbjörnsson, J. Kamps, and M. de Rijke: EuroGOV: Engineering a Multilingual Web
Corpus, In Proceedings of CLEF 2005, 2005.
[10] B. Sigurbjörnsson, J. Kamps, and M. de Rijke: WebCLEF 2005: Cross-Lingual Web Re-
trieval, In Proceedings of CLEF 2005, 2005.
[11] TextCat: Language identification tool, http://odur.let.rug.nl/ vannord/TextCat/, 2005.
[12] M. Tovar, M. Carrillo, D. Pinto, H. Jimenez, Combining Keyword Identification Techniques,
Research on Computing Science Journal, ISSN 1665-9899, 2005.
[13] R. Urbizagástegui: Las posibilidades de la Ley de Zipf en la indización automática, Research
report of the California Riverside University, 1999.
[14] G. K. Zipf: Human Behavior and the Principle of Least-Effort, Addison-Wesley, Cambridge
MA, 1949.