=Paper= {{Paper |id=Vol-1172/CLEF2006wn-GeoCLEF-LevelingEt2006 |storemode=property |title=University of Hagen at GeoCLEF2006: Experiments with Metonymy Recognition in Documents |pdfUrl=https://ceur-ws.org/Vol-1172/CLEF2006wn-GeoCLEF-LevelingEt2006.pdf |volume=Vol-1172 |dblpUrl=https://dblp.org/rec/conf/clef/LevelingV06a }} ==University of Hagen at GeoCLEF2006: Experiments with Metonymy Recognition in Documents== https://ceur-ws.org/Vol-1172/CLEF2006wn-GeoCLEF-LevelingEt2006.pdf
        University of Hagen at GeoCLEF 2006:
  Experiments with metonymy recognition in documents
                                   Johannes Leveling and Dirk Veiel
                            FernUniversität in Hagen (University of Hagen)
                     Intelligent Information and Communication Systems (IICS)
                                       58084 Hagen, Germany
                    {johannes.leveling, dirk.veiel}@fernuni-hagen.de


                                                Abstract
     This paper describes the participation of the IICS group at the GeoCLEF task of the CLEF
     campaign 2006. We describe different retrieval experiments using a separate index for location
     names and identifying and indexing of metonymic location names differently. The setup of our
     GIR system is a modified variant of the setup for GeoCLEF 2005.
         We apply a classifier for the identification of metonymic location names for preprocessing
     the documents. This classifier is based on shallow features only and was trained on manually
     annotated data from the German CoNLL-2003 Shared Task corpus for Language-Independent
     Named Entity Recognition and from a subset of the GeoCLEF newspaper corpus. After pre-
     processing, documents contain additional information for location names that are to be indexed
     separately, i.e. LOC (all location names identified), LOCLIT (location names in their literal
     sense), and LOCMET (location names in their metonymic sense).
         To obtain an IR query from the topic title, description, and narrative, we employ two meth-
     ods. In the first method, a semantic parser analyzes the query text and the resulting semantic
     net is transformed into database query. The second method uses a Boolean combination of a
     bag-of-words (consisting of topical search terms) with location names.
         The results of our experiments can be summarized as follows: excluding metonymic senses
     of location names improves mean average precision (MAP) for most of our experiments. For
     experiments in which this was not the case, a more detailed analysis showed that for some
     topics the precision increased. Our experiments show that the additional use of topic narratives
     decreases MAP. For almost all experiments with the topic narrative, lower values for MAP and
     for the number of relevant and retrieved documents were observed. However, query expansion
     and the use of separate indexes improves the performance of our GIR application.

Categories and Subject Descriptors
H.3.1 [Information Storage and Retrieval]: Content Analysis and Indexing—Indexing methods; Linguis-
tic processing; H.3.3 [Information Storage and Retrieval]: Information Search and Retrieval—Query
formulation; Search process; H.3.4 [Information Storage and Retrieval]: Systems and Software—Per-
formance evaluation (efficiency and effectiveness)

General Terms
Measurement, Performance, Experimentation

Keywords
Geographic Information Retrieval, Metonymy Recognition, Indexing Location Names
1    Introduction
There are several essential tasks a geographic information retrieval (GIR) application has to perform, for
example the identification and disambiguation of location names (see [3] for a general overview over tasks
in GIR applications). Disambiguating location names includes differentiating between their literal (geo-
graphic) and metonymic senses. Metonymy is typically defined as a figure of speech in which a speaker
uses “one entity to refer to another that is related to it” [4].
    This paper presents an application of a classifier for literal and metonymic senses of location names.
The classifier is trained on manually annotated data and uses shallow features from the textual context of
location names only. We create different indexes corresponding to different senses of location names and
investigate in a baseline experiment if utilizing a separate index containing location names will improve
performance in GIR. The focus in our experiments lies on metonym identification in documents, because
the GeoCLEF query topics did not contain any metonymic location names in topic titles, descriptions, or
narrative.


2    System description
Our GIR system is based on the same system that was developed for GeoCLEF 2005 (see [8]). The
WOCADI parser (Word Class Controlled Disambiguating (Parser), see [2]) analyzes the query topics and
the GeoCLEF corpus of newspaper and newswire articles. From its parse results, concepts (or rather:
lemmata) and compound constituents are extracted as index terms or search terms.
    For the identification of metonymic location names, we employed a classifier trained on manually an-
notated data. The data consists of a subset of the German CoNLL-2003 Shared Task corpus for Language-
Independent Named Entity Recognition [9] and a subset of the GeoCLEF newspaper corpus. The metonymy
classifier [7] is based on shallow features only (e.g., part-of-speech information for closed word classes
from list lookup, position of words in a sentence, word length, and base forms of verbs) and achieved a per-
formance of 81.7% F1 -measure in differentiating between literal and metonymic senses of location names.
In analyzing the annotated CoNLL data (1216 instances), we found that 16.95% of all location names were
used metonymically, and 7.73% referenced both a literal and a metonymic sense at the same time (see [7]
for a more detailed description of the metonymy classification). These numbers provide an upper bound for
a performance increase for methods exploiting metonymy information. After preprocessing, the documents
are structured with the following fields:
    • DOCID – document ID
    • TEXT – text of the document
    • LOC – location names from the text
    • LOCLIT – location names in their literal sense
    • LOCMET – location names in their metonymic sense
All identified location names are indexed from the LOC field of a document. The result of the metonymy
classifier determines how a given location name will be indexed, i.e. literal and metonymic senses of lo-
cation names are indexed from the LOCLIT and LOCMET fields, respectively. Figure 1 shows an example
document after preprocessing its text. The representations of 276,581 documents (after duplicate elimina-
tion) were indexed with the Zebra database management system [1], which supports a standard relevance
ranking (tf-idf IR model).
     Two methods were employed to obtain an IR query from a topic title, description, and narrative: In the
first method, the WOCADI parser is applied to perform a deep linguistic analysis of the query text and the
resulting semantic net is transformed into a database independent query representation (DIQR, see [6]). In
the second method, a bag-of-words (topical search terms extracted from the query text) is combined with
a subquery containing location names (identified by a name lookup) to obtain the DIQR. In both cases,
a DIQR query consists of a Boolean combination of a subquery containing of topical search terms (or
descriptors) and one containing location names.
                             
                              ... 
                             At the meeting of France and Germany
                             in Lisbon last year, Paris vetoed the decision.
                             ... 
                             France Germany Lisbon Paris
                             Lisbon
                             France Germany Paris
                             

Figure 1: An example of a preprocessed GeoCLEF document. Note: In this sample document, the TEXT
field contains the unprocessed text, before stemming and stopword removal.


    In our baseline experiment (FUHddGNNNTD), the standard IR model (tf-idf ) was utilized without
additions or modifications. In experiments not accessing the separate name index, the location names
were searched within the index for the TEXT field; in experiments using the separate index (e.g. FUHd-
dGYYYTDN) location names were looked for in the index for the LOC field. For experiments with
metonymy identification, the index for location names corresponds to the field LOCLIT, i.e. only loca-
tion names in their literal sense were searched for.


3      Description of the runs submitted and results
The parameters used in our GeoCLEF experiments can be described as follows:

     • BOP: Boolean operator for the combination of topic and location subquery

     • LI: use separate location index (LOC or LOCLIT, as described in Section 2) vs. use TEXT index

     • LA: apply deep linguistic analysis (WOCADI parser [2]) and transformation of resulting semantic
       net into a database query [6] vs. bag-of-words IR

     • QEX: query expansion with semantically related terms and meronyms for locations [8] vs. no ex-
       pansion

     • MET: exploit metonymy information (as described in [7]) vs. no metonymy processing

     • QF: query topic fields (TD: title and description, TDN: title, description, and narrative). A run with
       topic title, description, and narrative was mandatory to find out if extra narrative terms help improve
       performance.

    Table 1 and Table 2 show the different parameter settings and results for our monolingual German
and bilingual English-German GeoCLEF runs, respectively. The results shown consist of mean average
precision (MAP) and the number of relevant and retrieved documents (rel ret).


4      Analysis of the results
Using a separate index for location names leads to a better performance in our experiments. Additional
experiments confirmed that the query expansion and the provided background knowledge have significant
influence on results. As a preparation for the GeoCLEF 2006 experiments, we analyzed data from Geo-
CLEF 2005, in particular (but not exclusively) those topics for which we did not find any of the relevant
documents. We modified the knowledge base containing our background knowledge to include some miss-
ing facts1 and performed manual searching. For example, for topic GC019 (“European golf tournaments”),
    1 For the GeoCLEF 2006 experiments, the unmodified knowledge base from GeoCLEF 2005 was used.
Table 1: Parameter settings and results for monolingual German GeoCLEF experiments. 785 documents
were assessed as relevant for the 25 queries. (Additional runs are set in italics.)

          Run Identifier                            Parameters                         Results
                                    BOP      LI   LA     QEX     MET       QF       MAP     rel ret
          FUHddGNNNTD                OR      N     N       N       N       TD     0.1694         439
          FUHddGYYYTD                OR      Y     Y       Y       N       TD     0.2229         449
          FUHddGYYNTD                OR      Y     Y       N       N       TD     0.1865         456
          FUHddGNNNTDN               OR      N     N       N       N      TDN     0.1223         426
          FUHddGYYYTDN               OR      Y     Y       Y       N      TDN     0.2141         462
          FUHddGYYYMTDN              OR      Y     Y       Y       Y      TDN     0.1999         442
          FUHddGYYNTD               AND      Y     Y       N       N       TD     0.1466         232
          FUHddGYYNMTD              AND      Y     Y       N       Y       TD     0.1608         225
          FUHddGYYYTD               AND      Y     Y       Y       N       TD     0.1718         267
          FUHddGYYYMTD              AND      Y     Y       Y       Y       TD     0.1953         259


Table 2: Parameter settings for bilingual English-German GeoCLEF experiments. 785 documents were
assessed as relevant for the 25 queries.
          Run Identifier                          Parameters                 Results
                                    BOP     LI    LA    QEX      MET      QF       MAP      rel ret
          FUHedGNNNTD                OR      N     N      N        N      TD      0.1211         397
          FUHedGNNNTDN               OR      N     N      N        N     TDN      0.0548         333
          FUHedGYYYTD                OR      Y     Y      Y        N      TD      0.1280         383
          FUHedGYYNTD                OR      Y     Y      N        N      TD      0.1124         386
          FUHedGYYYTDN               OR      Y     Y      Y        N     TDN      0.1234         375
          FUHedGYYYMTDN              OR      Y     Y      Y        Y     TDN      0.1148         375


we found that with only slight differences in query formulation the MAP as well as the number of relevant
documents increased. For topic GC019, 41 of 61 relevant documents were found after modifying the query.
As there is now more data (topics and relevance assessments) available, we hope to identify regularities for
search failures more easily.
    As we observed from results of the GeoCLEF 2005 experiments, query expansion with meronyms
leads to significantly better precision (0.1466 MAP vs. 0.1608 MAP; 0.1718 MAP vs. 0.1953 MAP)
for most monolingual German runs, although recall is slightly worse. A detailed analysis of the runs
FUHddGYYYTDN an FUHddGYYYMTDN shows that in this case, the metonymy identification task
added this year improves the results for some topics.
    Results for the bilingual experiments were found to be lower in general. There were a few errors in the
translated topic titles, descriptions, and narratives. For some topic titles, there does not seem to be enough
textual context to provide an adequate translation (topic titles and descriptions were translated separately).
    One hypothesis to be tested was that the additional information in topic narratives (runs with QEX=TDN
instead of TD) would improve results. We can not confirm this assumption because with our setup, MAP
and relevant and retrieved documents are almost always lower for runs using the topic narrative than for
runs with topic title and description.


5    Conclusion
We expected to find a significant increase in precision for all GIR experiments excluding metonymic senses
of location names for a search. This assumption holds for most experiments, conforming results of earlier
experiments [7]. However, the MAP for experiments using metonymy information is in one case lower. A
more detailed analysis of results for this experiment showed that at least for some topics, precision is in
fact increased by metonymy identification.
    A different explanation for a low performance might be that in our setup for GeoCLEF 2006, the loca-
tion name index does not contain terms representing adjectives, language, or inhabitants for a given location
name (e.g. the terms like Dutch, Spanish, or Spaniard do not occur in the location name index). Further-
more, instead of removing all metonymic senses for location names from the index, the corresponding
terms should be indexed differently (with a lesser weight or with a different sense).
    Additional information from the topic narratives did not improve precision or recall in our experiments,
although one might have expected a similar effect as for query expansion with meronyms. We plan to rerun
experiments with a state-of-the-art database management system (DBMS), such as Cheshire 3 [5], which
offers a variety of different IR models. A more modern IR model will help to increase performance in
general.


References
[1] Sebastian Hammer, Adam Dickmeiss, Heikki Levanto, and Mike Taylor. Zebra – user’s guide and
    reference. Manual, IndexData, Copenhagen, Denmark, 2005.

[2] Sven Hartrumpf. Hybrid Disambiguation in Natural Language Analysis. Der Andere Verlag, Os-
    nabrück, Germany, 2003.

[3] Christopher B. Jones, R. Purves, Anne Ruas, Mark Sanderson, Monika Sester, Marc J. van Kreveld,
    and Robert Weibel. Spatial information retrieval and geographical ontologies – an overview of the
    SPIRIT project. In SIGIR 2002, pages 387–388, 2002.

[4] George Lakoff and Mark Johnson. Metaphors we live by. Chicago University Press, 1980.

[5] Ray R. Larson and Robert Sanderson. Grid-based digital libraries: Cheshire 3 and distributed retrieval.
    In JCDL ’05: Proceedings of the 5th ACM/IEEE-CS Joint Conference on Digital Libraries, pages
    112–113, New York, NY, USA, 2005. ACM Press.

[6] Johannes Leveling. Formale Interpretation von Nutzeranfragen für natürlichsprachliche Interfaces zu
    Informationsangeboten im Internet. Dissertation, Fachbereich Informatik, FernUniversität in Hagen,
    2006. To appear in: Der andere Verlag, Tönning, Germany, 2006.

[7] Johannes Leveling and Sven Hartrumpf. On metonymy recognition for GIR. In Proceedings of the 3rd
    ACM workshop on geographic information retrieval, 2006.

[8] Johannes Leveling, Sven Hartrumpf, and Dirk Veiel. Using semantic networks for geographic informa-
    tion retrieval. In C. Peters, F. C. Gey, J. Gonzalo, H. Müller, G. J. F. Jones, M. Kluck, B. Magnini, and
    M. de Rijke, editors, CLEF 2005 Proceedings, Lecture Notes in Computer Science (LNCS). Springer,
    Berlin, 2006.

[9] E. Tjong Kim Sang and F. De Meulder. Introduction to the CoNLL-2003 shared task: Language
    independent named entity recognition. In W. Daelemans and M. Osborne, editors, Proceedings of
    CoNLL-2003, pages 142–147, Edmonton, Canada, 2003.