<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>University of Hagen at CLEF 2005: Towards a Better Baseline for NLP Methods in Domain-Specific Information Retrieval</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Johannes Leveling</string-name>
          <email>Johannes.Leveling@fernuni-hagen.de</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>General Terms</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>58084 Hagen</institution>
          ,
          <country country="DE">Germany</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Hagen, FernUniversita ̈t in Hagen</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>The third participation of the University of Hagen at the German Indexing and Retrieval Test (GIRT) task of the Cross Language Evaluation Campaign (CLEF 2005) aims at providing a better baseline for experiments with natural language processing (NLP) methods in domainspecific information retrieval (IR). Our monolingual experiments with the German document collection are based on a setup combining several methods to achieve a better performance. The setup includes an entry vocabulary module (EVM), query expansion with semantically related concepts, and a blind feedback technique. The monolingual experiments focus on comparing two techniques for constructing database queries: creating a 'bag of words' and creating a semantic network by means of a syntactico-semantic parser for a deep linguistic analysis of the query. The best performance in the official experiments was achieved by a setup using staged logistic regression, a query expansion with semantically related concepts, an entry vocabulary module, a deep linguistic analysis of the query, and blind feedback (0.2875 mean average precision (MAP)). Additional experiments showed a performance improvement when changing to the basic Okapi BM25 search (0.3878 MAP). For the bilingual experiments, the English topics are translated into German queries with several machine translation services available online (Systran, Free translation, WorldLingo, and Promt). Each set of translated topics is processed separately with the same techniques as in the monolingual experiments. The best performance was achieved with a query translation by Promt with a simple keyword extraction from the translation (0.2399 MAP with a staged logistic regression approach vs. 0.2807 MAP with Okapi BM25).</p>
      </abstract>
      <kwd-group>
        <kwd>H</kwd>
        <kwd>3</kwd>
        <kwd>1 [Information Storage and Retrieval]</kwd>
        <kwd>Content Analysis and Indexing</kwd>
        <kwd>Indexing methods</kwd>
        <kwd>Linguistic processing</kwd>
        <kwd>H</kwd>
        <kwd>3</kwd>
        <kwd>3 [Information Storage and Retrieval]</kwd>
        <kwd>Information Search and Retrieval</kwd>
        <kwd>Query formulation</kwd>
        <kwd>Search process</kwd>
        <kwd>H</kwd>
        <kwd>3</kwd>
        <kwd>4 [Information Storage and Retrieval]</kwd>
        <kwd>Systems and Software</kwd>
        <kwd>Performance evaluation (efficiency and effectiveness)</kwd>
        <kwd>I</kwd>
        <kwd>2</kwd>
        <kwd>4 [Artificial Intelligence]</kwd>
        <kwd>Knowledge Representation Formalisms and Methods</kwd>
        <kwd>Semantic networks</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>1</p>
    </sec>
    <sec id="sec-2">
      <title>Introduction</title>
      <p>
        Natural language processing for information retrieval, Deep linguistic analysis, Keyword extraction
This paper presents the work for the third participation of the University of Hagen in the domain-specific
GIRT task in the evaluation campaign of the Cross Language Evaluation Forum (CLEF). Natural language
processing (NLP) as described in the following subsections is part of query processing for the NLI-Z39.501
        <xref ref-type="bibr" rid="ref15">(Leveling and Helbig, 2002)</xref>
        , a natural language interface for databases supporting the Internet protocol
Z39.50
        <xref ref-type="bibr" rid="ref10">(ISO, 1998)</xref>
        . A major part of our experimental infrastructure was developed for and is applied in
this NLP system.
      </p>
      <p>
        In CLEF 2003, retrieval strategies based on generating query variants for a single natural language (NL)
topic were compared
        <xref ref-type="bibr" rid="ref13 ref14">(Leveling, 2004)</xref>
        . The best experiment of the University of Hagen in the
domainspecific task in 2003 with respect to mean average precision (MAP) used multiple variants of a query, a
query expansion with semantically related concepts, and a database index containing word forms (0.2064
MAP for a run using both topic title and description).
      </p>
      <p>
        In CLEF 2004, the focus of our experiments was on investigating differences in indexing methods, such
as indexing unprocessed word forms, indexing concepts, and indexing semantic networks
        <xref ref-type="bibr" rid="ref14 ref16 ref9">(Leveling and
Hartrumpf, 2005)</xref>
        . The best experiment of the University of Hagen in the domain-specific task in 2004 with
respect to mean average precision used a single query, query expansion with semantically related concepts,
and a database index containing word forms (0.2482 MAP for a monolingual German run using both topic
title and description).
      </p>
      <p>For the monolingual experiments in CLEF 2005 two main objectives are pursued:
1. To establish a better baseline for a comparison between NLP methods and traditional approaches
in IR and to achieve a better overall performance. To complete this goal, some methods from the
experimental setup of the UC Berkeley in previous years were adapted.
2. To compare two techniques for creating database queries from the natural language topics: a)
extracting keywords (’bag of words’) and b) applying a deep linguistic analysis by means of a
syntacticosemantic parser before creating a database query.</p>
      <p>For the bilingual experiments with the GIRT collection, the English query topics are translated into
German by means of several free machine translation services that are available over the Internet. The
translations obtained from each machine translation service are processed in separate experiments,
employing the query processing techniques for monolingual experiments described above.
1.1</p>
      <sec id="sec-2-1">
        <title>Towards a Better Baseline</title>
        <p>To state and investigate a hypothesis such as “The use of semantic networks (or NLP in general) – including
methods such as expanding queries with semantically related concepts – improves performance in IR.” an
acceptable experimental baseline has to be established.</p>
        <p>The experimental setup used by the University of Hagen in previous participations at CLEF did not
provide a performance higher than other state-of-the-art systems. It served mainly as a basis for comparing
different strategies on the same system (intra-system comparison) and did not typically aim at an overall
superior performance (inter-system comparison).</p>
        <p>To obtain a better baseline for our experiments, several methods of the experiments of the UC Berkeley
for CLEF were added to our experimental setup to improve retrieval performance, including a so-called
entry vocabulary module. In total, our setup now includes the following options:</p>
        <p>1The NLI-Z39.50 is being developed as part of the project “Natu¨rlichsprachliches Interface fu¨r die internationale
Standardschnittstelle Z39.50” and funded by the DFG (Deutsche Forschungsgemeinschaft) within the support program for libraries
“Modernisierung und Rationalisierung in wissenschaftlichen Bibliotheken”.</p>
        <p>
          • LA: extracting keywords (’bag of words’) or applying a deep linguistic analysis, i.e. employing NLP
methods to produce a semantic network representation (described in
          <xref ref-type="bibr" rid="ref13 ref14 ref14 ref16 ref9">(Leveling, 2004; Leveling and
Hartrumpf, 2005)</xref>
          ).
• EVM: employing an entry vocabulary module. An entry vocabulary module provides a mapping of
words from a possibly uncontrolled vocabulary to a controlled vocabulary, based on likelihoods of
co-occurrence
          <xref ref-type="bibr" rid="ref11 ref4 ref5 ref6">(Gey and Chen, 1997; Gey et al., 2001b,a)</xref>
          . As suggested in
          <xref ref-type="bibr" rid="ref16">(Petras, 2005)</xref>
          , the top five
ranked terms from the EVM are added to the database query, down weighted by half of the default
weight for our experimental runs.
• BF: using blind feedback. The top-N controlled terms from the top-M ranked documents are
extracted for a query reformulation
          <xref ref-type="bibr" rid="ref16 ref16 ref17">(Petras et al., 2004; Petras, 2005)</xref>
          . As suggested in
          <xref ref-type="bibr" rid="ref16 ref17">(Petras et al.,
2004)</xref>
          , thirty terms from the top twenty documents are extracted and down weighted by half of default
weight in our experiments (N = 30, M = 20).
• QEX: adding semantically related concepts to the query. The query expansion stage is based on
adding semantically related concepts (word senses) to a query, including synonyms, hyponyms,
meronyms. This approach was described in
          <xref ref-type="bibr" rid="ref13 ref14">(Leveling, 2004)</xref>
          .
        </p>
        <sec id="sec-2-1-1">
          <title>In addition, our database setup consists of:</title>
          <p>
            • The Cheshire II database
            <xref ref-type="bibr" rid="ref11 ref12 ref5">(Larson and McDonough, 1997; Larson et al., 1996)</xref>
            , which supports
Boolean searches, probabilistic weighting with staged logistic regression
            <xref ref-type="bibr" rid="ref2 ref3">(Cooper et al., 1992, 1994)</xref>
            ,
which is used in the UC Berkeley setup as a default, and basic Okapi BM25
            <xref ref-type="bibr" rid="ref18">(Robertson et al., 1994)</xref>
            (opposed to the Zebra database used in our earlier experiments; Zebra supports Boolean searches
and the standard tf-idf weighting as a default).
• A document representation which results from applying
– the WOCADI parser (WOrd ClAss based DIsambiguating parser) to obtain results for the
morpho-lexical analysis for the title and abstract
            <xref ref-type="bibr" rid="ref7">(opposed to indexing unprocessed word forms
as in our experiments for CLEF 2003)</xref>
            ,
– a stopword list of a few hundred word forms (opposed to indexing all word forms, concepts, or
semantic networks as in our experiments for CLEF 2004),
– a German stemmer, which was originally implemented in Snowball2 (opposed to indexing word
forms or lemmata), and
– a linguistically motivated (lexicon-based) decomposition of German noun compounds.
1.2
          </p>
        </sec>
      </sec>
      <sec id="sec-2-2">
        <title>Techniques for Query Processing</title>
        <p>
          Two techniques for query processing are compared. The first technique corresponds to extracting keywords
from the topic title and topic description to create a database query. The query string is tokenized and word
forms are extracted. Some normalization steps such as stopword removal and stemming are employed
to produce a database query in the Database Independent Query Representation (DIQR, see
          <xref ref-type="bibr" rid="ref13 ref14">(Leveling,
2004)</xref>
          ). This roughly corresponds to the traditional approach of ‘processing’ natural language queries in
information retrieval (’bag of words’).
        </p>
        <p>
          The second query processing technique employs NLP methods to create a database query. A
syntacticosemantic parser, WOCADI
          <xref ref-type="bibr" rid="ref7">(Hartrumpf, 2003)</xref>
          , produces a semantic network representation of the query
according to the MultiNet paradigm
          <xref ref-type="bibr" rid="ref8 ref9">(Helbig, 2001, 2005)</xref>
          which is then transformed into a DIQR by means
of a rule-based transformation engine consisting of a Rete compiler and a Rete interpreter (the
implementation is described in more detail in
          <xref ref-type="bibr" rid="ref15">(Leveling and Helbig, 2002)</xref>
          ).
        </p>
        <p>For both techniques, the DIQR is mapped to a query in a formal language the database management
software supports (such as a query for the Z39.50 protocol) which is then submitted to the target database.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Monolingual GIRT Experiments (German – German)</title>
      <sec id="sec-3-1">
        <title>Experimental Setup and Results</title>
        <p>For the GIRT task in 2005, several experimental runs for the monolingual GIRT task were submitted. The
experiments vary in the following parameter settings: using a query expansion with semantically related
terms (QEX=yes/no), using an entry vocabulary module (EVM=yes/no), constructing a query from the
semantic network after a linguistic analysis or using a traditional keyword extraction (LA=yes/no), and
using blind feedback (BF=yes/no). Table 1 gives an overview of the monolingual experiments performed
for the German document collection with their results.
2.2</p>
      </sec>
      <sec id="sec-3-2">
        <title>Discussion of Results</title>
        <p>The best monolingual GIRT experiment used a query expansion with semantically related concepts, an
entry vocabulary module, and blind feedback. The performance of the best official experiment with respect
to mean average precision (0.3031 MAP for run FUHggyydbf) is better in comparison to our experiments
in CLEF 2003 (0.2064 MAP) and CLEF 2004 (0.2482 MAP). Additional experiments were conducted
using the state-of-the-art weighting approach, Okapi BM25, which yield a significantly better performance
(0.3878 MAP for run FUHggyylbf).</p>
        <p>The effect of any single query processing method (corresponding to a single parameter) is still
inconclusive, but the combination of all processing methods with a deep linguistic analysis of the query yields
the best performance with respect to the number of relevant and retrieved documents and mean average
precision (FUHggyylbf).</p>
        <p>Several differences between our setup and the experimental setup of the UC Berkeley still remain and
may account for a difference in performance.</p>
        <p>•</p>
        <sec id="sec-3-2-1">
          <title>Morphological processing:</title>
          <p>In the setup of the UC Berkeley, the stemmer is applied to word forms. In our setup, stemming is
applied after lemmatizing. The WOCADI parser returns morpho-lexical results for the sentences,
including the lemmata (complex named entities and multi-word lexemes are not identified in the
morpho-lexical analysis at the moment). The input data for the German stemmer consists of these
lemmata from the morpho-lexical analysis of a sentence, i.e. the stemming process starts with
normalized word forms. Plural forms of nouns, participles, past forms of verbs and verb with a separable
prefix are already normalized.
• Database Index:</p>
          <p>In the setup of the UC Berkeley, only the constituents of German compounds are used as index
terms. In our setup, compounds are indexed together with their constituents (because of a later query
expansion with semantically related compounds). In addition, our database index contains stemmed
word forms from the title and the abstract fields of the documents only; a database index containing
subject terms as well was not created.
• Phrases:</p>
          <p>In our setup, a simple frequency-based phrase recognition for the EVM is employed which identifies
adjective-noun sequences (English noun phrases often correspond to compounds in German).
• Compounds:</p>
          <p>
            In the UC Berkeley setup, decompounding of compounds is based on a statistical approach
            <xref ref-type="bibr" rid="ref1">(Chen,
2002)</xref>
            . In our setup, decompounding is lexicon-based and includes a solution of the so called
Fugen-problem (omitting, inserting, or substituting letters in or between constituents for a German
compound, as an omitted ‘e’ in ‘Schulsprecher’ – ‘Schule’ + ‘Sprecher’), an additional ‘s’
(‘Verfahrensfehler’ – ‘Verfahren’ + ‘Fehler’), or an Umlaut in the plural form (‘Ga¨nsefleisch’ – ‘Gans’
+ ‘Fleisch’).3 Compounds which should not be split into their constituents are entered into the
computational lexicon to block decompounding (e.g., ‘Frauenzimmer’/‘dame’ does not have regular
semantic relations to its constituents and is represented as a concept).
• Entry Vocabulary Module:
          </p>
          <p>In our setup, the EVM uses a co-occurrence between title words and subject terms only and does not
include words from the abstract.</p>
          <p>These differences can account for a significantly lower performance of our experiments. For example,
we conducted tests which showed a different performance for indexing lemmata and indexing stemmed
word forms. Experiments on the index with stemmed word forms resulted in a significantly better
performance because the stemmed forms conflate adjectives and nouns into the same form (e.g.
‘wirtschaftlich’/‘economical’ and ‘Wirtschaft’/‘economy’ are conflated into ‘wirtschaft’/‘economi’). For a database
index with lemmata, only inflectional endings are removed and e.g. ‘wirtschaftlich’ and ‘wirtschaft’
remain different index terms. While treating different concepts differently remains one of our objectives
for a high precision information search, our background knowledge which semantically links the correct
senses of ‘Wirtschaft’ and ‘wirtschaftlich’ is neither robust (i.e., it relies on correct word sense
disambiguation) nor lexically complete, i.e. the linking between corresponding concepts (derivational links) in our
computational lexicon is not complete, yet.
3</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Bilingual GIRT Experiments (English – German)</title>
      <p>In CLEF 2004, the bilingual experiments of the University of Hagen were based on a method which
combines a concept translation using linguistic resources (such as GermaNet and EuroWordNet) with a word
translation using translation lists to obtain a ranked list of translation alternatives. This method was
employed to translate German concepts in the DIQR into English to create a database query. For CLEF 2005,
English query topics are submitted to freely available machine translation services for a translation into
German queries.
3.1</p>
      <sec id="sec-4-1">
        <title>Experimental Setup and Results</title>
        <p>Our second participation in the bilingual GIRT task (matching English topics against the German data)
is based on various machine translation services for a translation. For the bilingual retrieval experiments
3A ‘Fuge’ does not correspond to a morphological suffix and can not be treated by a suffix elimination process (stemming).
(English – German) with the GIRT document collection, four machine translation services translate the
English topics into German queries: Free translation4, Systran5, and WorldLingo6, and Promt7.</p>
        <p>For the official runs, keywords are extracted from the translated queries: the query is tokenized,
stopwords are removed, and a stemmer is applied. The remaining stemmed words are employed to create
a simple DIQR, which is processed accordingly. For a fifth experiment, the topics are translated by the
Promt translator, analyzed by WOCADI to produce a semantic network representation, and transformed
into a DIQR with the Rete-based transformation. The DIQR queries are then processed as described in
Section 2. Table 2 gives an overview of the bilingual experiments with English topics on the German
document collection together with their results.
3.2</p>
      </sec>
      <sec id="sec-4-2">
        <title>Discussion of Results</title>
        <p>The WOCADI parser may serve as an indicator for the quality of the translated queries. WOCADI performs
many morpho-syntactical and semantical tests while trying to produce a semantic network representation
for a query topic. If some of these tests (such as agreement between subject and verb, selectional restrictions
for the complement of an action (verb), etc.) fail or if the translation contains other errors, a partial semantic
network is returned. For severe errors in the translations, no network is returned at all.</p>
        <p>WOCADI reported no problems in parsing the original German topics. Parse results for the translated
German queries showed that many of the 200 translations (25 topics, four different machine translations
services, and translation of title and description) contained a syntactical or lexical error, so that WOCADI
could not fully analyze these translations. For example, all services had problems translating the imperative
forms of the verb ‘find’. (The exclamation mark to indicate the imperative is missing in all topics and the
original queries were not modified by us.)</p>
        <p>Due to these test results, we judge the quality of online translations so poor that the results from
translating English topics into German can not be processed successfully with a deep linguistic analysis. Thus,
the effect of trying to parse a grammatically incorrect translation (a partial semantic network or no
semantic network is returned by WOCADI) is not present in a simple keyword extraction from a shallow
analysis (the morpho-lexical stage in the WOCADI parser). All potential advantages of a deep linguistic
processing are then unavailable (FUHegpyyn vs. FUHegpyyl). A performance increase was observed for
all experiments using Okapi BM25.</p>
        <p>4http://www.freetranslation.com/
5http://www.systransoft.com/
6http://www.worldlingo.com/
7http://www.e-promt.com/en/</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusion</title>
      <p>
        In comparison with the results for the monolingual GIRT task in 2003 and 2004, performance with respect
to the MAP for the best official experiment has improved considerably
        <xref ref-type="bibr" rid="ref7">(0.2064 MAP in 2003, 0.2482 in
2004 vs. 0.3031 in 2005)</xref>
        . Additional experiments involved changing the ranking scheme to Okapi BM25,
which increased the number of relevant and retrieved documents and the mean average precision
significantly (0.3878 MAP). As indicated by the better performance, the setup for the additional experiments
provides a much better baseline for experiments with NLP methods in IR.
      </p>
      <p>The method for constructing a database query using the transformation of the semantic network
representation into a database query yields a higher performance than extracting keywords in combination with
all other methods applied in these experiments. Results are still inconclusive in which cases NLP methods
provide a better performance and even seem to depend on the ranking scheme employed.</p>
      <p>The machine translation services tested did not produce high-quality translations. At the moment,
using a robust keyword extraction yields better performance than a subsequent try to semantically analyze
malformed ‘translations’.</p>
      <p>The experimental setup of the University of Hagen and of the UC Berkeley still differ much, despite
using methods which are comparable, similar or even the same with respect to some experimental parameters
for the monolingual experiments. Further research is required and should properly identify the differences
in these parameters and their influence on the overall retrieval performance.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          <string-name>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <surname>Aitao</surname>
          </string-name>
          (
          <year>2002</year>
          ).
          <article-title>Multilingual information retrieval using English and Chinese queries</article-title>
          .
          <source>In Evaluation of Cross-Language Information Retrieval Systems</source>
          , Second Workshop of the Cross-
          <article-title>Language Evaluation Forum, CLEF 2001 (edited by</article-title>
          <string-name>
            <surname>Peters</surname>
          </string-name>
          , C.;
          <string-name>
            <surname>M. Braschler; J. Gonzalo</surname>
          </string-name>
          ; and M. Kluck), volume
          <volume>2406</volume>
          of Lecture Notes in Computer Science (LNCS), pp.
          <fpage>44</fpage>
          -
          <lpage>58</lpage>
          . Berlin: Springer.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          <string-name>
            <surname>Cooper</surname>
          </string-name>
          , William S.;
          <string-name>
            <surname>Fredric C. Gey</surname>
            ; and
            <given-names>Aitoa</given-names>
          </string-name>
          <string-name>
            <surname>Chen</surname>
          </string-name>
          (
          <year>1994</year>
          ).
          <article-title>Probabilistic retrieval in the TIPSTER collections: An application of staged logistic regression</article-title>
          .
          <source>In Proceedings of the Second Text REtrieval Conference (TREC-2)</source>
          <article-title>(edited by</article-title>
          <string-name>
            <surname>Harman</surname>
          </string-name>
          , D.), pp.
          <fpage>57</fpage>
          -
          <lpage>66</lpage>
          . National Institute of Standards and Technology (NIST),
          <source>Special Publication 500-215.</source>
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          <string-name>
            <surname>Cooper</surname>
          </string-name>
          , William S.;
          <string-name>
            <surname>Fredric C. Gey</surname>
          </string-name>
          ; and
          <string-name>
            <surname>Daniel P. Dabney</surname>
          </string-name>
          (
          <year>1992</year>
          ).
          <article-title>Probabilistic retrieval based on staged logistic regression</article-title>
          .
          <source>In SIGIR '92: Proceedings of the 15th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          , pp.
          <fpage>198</fpage>
          -
          <lpage>210</lpage>
          . New York, NY, USA: ACM Press.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          <string-name>
            <surname>Gey</surname>
          </string-name>
          , Fredric C.
          <article-title>; Michael Buckland; Aitao Chen;</article-title>
          and
          <string-name>
            <surname>Ray R. Larson</surname>
          </string-name>
          (
          <year>2001a</year>
          ).
          <article-title>Entry vocabulary - a technology to enhance digital search</article-title>
          .
          <source>In Proceedings of the First International Conference on Human Language Technology</source>
          . San Diego.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          <string-name>
            <surname>Gey</surname>
            ,
            <given-names>Fredric C.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Aitao Chen</surname>
          </string-name>
          (
          <year>1997</year>
          ).
          <article-title>Phrase discovery for English and cross-language retrieval at TREC 6</article-title>
          .
          <source>In Proceedings of the Sixth Text REtrieval Conference (TREC-6)</source>
          <article-title>(edited by</article-title>
          <string-name>
            <surname>Voorhees</surname>
            , E. M. and
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Harman</surname>
          </string-name>
          ), pp.
          <fpage>637</fpage>
          -
          <lpage>647</lpage>
          . National Institute of Standards and Technology (NIST),
          <source>Special Publication 500-240.</source>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <string-name>
            <surname>Gey</surname>
          </string-name>
          , Fredric C.
          <article-title>; Hailing Jiang; Vivien Petras;</article-title>
          and
          <string-name>
            <given-names>Aitao</given-names>
            <surname>Chen</surname>
          </string-name>
          (
          <year>2001b</year>
          ).
          <article-title>Cross-language retrieval for the CLEF collections - comparing multiple methods of retrieval</article-title>
          .
          <source>In Cross-Language Information Retrieval and Evaluation: Workshop of Cross-Language Evaluation Forum</source>
          ,
          <string-name>
            <surname>CLEF</surname>
          </string-name>
          <year>2000</year>
          (
          <article-title>edited by</article-title>
          <string-name>
            <surname>Peters</surname>
          </string-name>
          , C.),
          <source>volume 2069 of Lecture Notes in Computer Science (LNCS)</source>
          , pp.
          <fpage>116</fpage>
          -
          <lpage>128</lpage>
          . Berlin: Springer.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <string-name>
            <surname>Hartrumpf</surname>
          </string-name>
          ,
          <string-name>
            <surname>Sven</surname>
          </string-name>
          (
          <year>2003</year>
          ).
          <article-title>Hybrid Disambiguation in Natural Language Analysis</article-title>
          .
          <source>Osnabru¨ck, Germany: Der Andere Verlag. ISBN 3-89959-080-5.</source>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <string-name>
            <surname>Helbig</surname>
          </string-name>
          , Hermann (
          <year>2001</year>
          ).
          <article-title>Die semantische Struktur natu¨rlicher Sprache: Wissensrepra¨sentation mit MultiNet</article-title>
          . Berlin: Springer.
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          <string-name>
            <surname>Helbig</surname>
          </string-name>
          , Hermann (
          <year>2005</year>
          ).
          <article-title>Knowledge Representation and the Semantics of Natural Language</article-title>
          . Berlin: Springer.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          <source>ISO (editor)</source>
          (
          <year>1998</year>
          ). ISO 23950:
          <year>1998</year>
          , Information and documentation - Information
          <source>retrieval (Z39</source>
          .50)
          <article-title>- Application service definition and protocol specification</article-title>
          . Geneva: International Organization for Standardization.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          <string-name>
            <surname>Larson</surname>
          </string-name>
          ,
          <string-name>
            <surname>Ray R</surname>
          </string-name>
          . and
          <string-name>
            <surname>Jerome McDonough</surname>
          </string-name>
          (
          <year>1997</year>
          ).
          <article-title>Cheshire II at TREC 6: Interactive probabilistic retrieval</article-title>
          .
          <source>In Proceedings of the Sixth Text REtrieval Conference (TREC-6)</source>
          <article-title>(edited by</article-title>
          <string-name>
            <surname>Voorhees</surname>
            , E. M. and
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Harman</surname>
          </string-name>
          ), pp.
          <fpage>649</fpage>
          -
          <lpage>660</lpage>
          . National Institute of Standards and Technology (NIST),
          <source>Special Publication 500-240.</source>
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          <string-name>
            <surname>Larson</surname>
          </string-name>
          , Ray R.;
          <string-name>
            <surname>Jerome McDonough; Paul O'Leary; Lucy Kuntz</surname>
          </string-name>
          ; and Ralph
          <string-name>
            <surname>Moon</surname>
          </string-name>
          (
          <year>1996</year>
          ).
          <article-title>Cheshire II: designing a next-generation online catalog</article-title>
          .
          <source>Journal of the American Society for Information Science</source>
          ,
          <volume>47</volume>
          (
          <issue>7</issue>
          ):
          <fpage>555</fpage>
          -
          <lpage>567</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          <string-name>
            <surname>Leveling</surname>
          </string-name>
          ,
          <string-name>
            <surname>Johannes</surname>
          </string-name>
          (
          <year>2004</year>
          ). University of Hagen at CLEF 2003:
          <article-title>Natural language access to the GIRT4 data</article-title>
          .
          <source>In Comparative Evaluation of Multilingual Information Access Systems, 4th Workshop of the Cross-Language Evaluation Forum</source>
          ,
          <string-name>
            <surname>CLEF</surname>
          </string-name>
          <year>2003</year>
          ,
          <article-title>Revised Selected Papers (edited by Peters</article-title>
          ,
          <source>Carol; Julio Gonzalo; Martin Braschler; and Michael Kluck)</source>
          , volume
          <volume>3237</volume>
          of Lecture Notes in Computer Science (LNCS), pp.
          <fpage>412</fpage>
          -
          <lpage>424</lpage>
          . Berlin: Springer.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          <string-name>
            <surname>Leveling</surname>
          </string-name>
          , Johannes and Sven
          <string-name>
            <surname>Hartrumpf</surname>
          </string-name>
          (
          <year>2005</year>
          ). University of Hagen at CLEF 2004:
          <article-title>Indexing and translating concepts for the GIRT task</article-title>
          .
          <source>In Multilingual Information Access for Text</source>
          ,
          <article-title>Speech and Images: Results of the Fifth CLEF Evaluation Campaign (edited by</article-title>
          <string-name>
            <surname>Peters</surname>
          </string-name>
          , C.;
          <string-name>
            <surname>P. D. Clough; G. J. F. Jones; J. Gonzalo; M. Kluck</surname>
            ; and
            <given-names>B. Magnini)</given-names>
          </string-name>
          , volume
          <volume>3491</volume>
          of Lecture Notes in Computer Science (LNCS), pp.
          <fpage>271</fpage>
          -
          <lpage>282</lpage>
          . Berlin: Springer.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          <string-name>
            <surname>Leveling</surname>
          </string-name>
          , Johannes and Hermann Helbig (
          <year>2002</year>
          ).
          <article-title>A robust natural language interface for access to bibliographic databases</article-title>
          .
          <source>In Proceedings of the 6th World Multiconference on Systemics, Cybernetics and Informatics (SCI</source>
          <year>2002</year>
          )
          <article-title>(edited by Callaos</article-title>
          ,
          <source>Nagib; Maurice Margenstern; and Belkis Sanchez)</source>
          , volume XI, pp.
          <fpage>133</fpage>
          -
          <lpage>138</lpage>
          . International Institute of Informatics and Systemics (IIIS), Orlando, Florida.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          <string-name>
            <surname>Petras</surname>
          </string-name>
          ,
          <string-name>
            <surname>Vivian</surname>
          </string-name>
          (
          <year>2005</year>
          ).
          <article-title>GIRT and the use of subject metadata for retrieval. In Multilingual Information Access for Text, Speech and Images: Results of the Fifth CLEF Evaluation Campaign</article-title>
          ,
          <string-name>
            <surname>CLEF</surname>
          </string-name>
          <year>2004</year>
          (
          <article-title>edited by</article-title>
          <string-name>
            <surname>Peters</surname>
          </string-name>
          , C.;
          <string-name>
            <surname>P. D. Clough; G. J. F. Jones; J. Gonzalo; M. Kluck</surname>
            ; and
            <given-names>B. Magnini)</given-names>
          </string-name>
          , volume
          <volume>3491</volume>
          of Lecture Notes in Computer Science (LNCS), pp.
          <fpage>219</fpage>
          -
          <lpage>225</lpage>
          . Berlin: Springer.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          <string-name>
            <surname>Petras</surname>
          </string-name>
          , Vivien; Natalia Perelman; and
          <string-name>
            <surname>Fredric C. Gey</surname>
          </string-name>
          (
          <year>2004</year>
          ). UC Berkeley at CLEF-2003
          <article-title>- Russian language experiments and domain-specific retrieval</article-title>
          .
          <source>In Comparative Evaluation of Multilingual Information Access Systems, 4th Workshop of the Cross-Language Evaluation Forum</source>
          ,
          <string-name>
            <surname>CLEF</surname>
          </string-name>
          <year>2003</year>
          (
          <article-title>edited by</article-title>
          <string-name>
            <surname>Peters</surname>
          </string-name>
          , C.;
          <string-name>
            <surname>J. Gonzalo; M. Braschler</surname>
          </string-name>
          ; and M. Kluck), volume
          <volume>3237</volume>
          of Lecture Notes in Computer Science (LNCS), pp.
          <fpage>401</fpage>
          -
          <lpage>411</lpage>
          . Berlin: Springer.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          <string-name>
            <surname>Robertson</surname>
          </string-name>
          , Stephen E.; Steve Walker; Susan Jones;
          <article-title>Micheline Hancock-Beaulieu;</article-title>
          and
          <string-name>
            <given-names>Mike</given-names>
            <surname>Gatford</surname>
          </string-name>
          (
          <year>1994</year>
          ).
          <article-title>Okapi at TREC-3</article-title>
          .
          <source>In Proceedings of the Third Text REtrieval Conference (TREC-3)</source>
          <article-title>(edited by</article-title>
          <string-name>
            <surname>Harman</surname>
          </string-name>
          , D.), pp.
          <fpage>109</fpage>
          -
          <lpage>126</lpage>
          . National Institute of Standards and Technology (NIST),
          <source>Special Publication 500-226.</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>