=Paper= {{Paper |id=Vol-1179/CLEF2013wn-CHiC-WilhelmSteinEt2013 |storemode=property |title=Identifying the Most Suitable Stemmer for the CHiC Multilingual Ad-hoc Task |pdfUrl=https://ceur-ws.org/Vol-1179/CLEF2013wn-CHiC-WilhelmSteinEt2013.pdf |volume=Vol-1179 |dblpUrl=https://dblp.org/rec/conf/clef/Wilhelm-SteinSE13 }} ==Identifying the Most Suitable Stemmer for the CHiC Multilingual Ad-hoc Task== https://ceur-ws.org/Vol-1179/CLEF2013wn-CHiC-WilhelmSteinEt2013.pdf
    Identifying the most suitable stemmer for the
            CHiC multilingual ad-hoc task

         Thomas Wilhelm-Stein, Benjamin Schürer, and Maximilian Eibl

             Technische Universität Chemnitz, 09107 Chemnitz, Germany,
                     {wilt, schben, eibl}@hrz.tu-chemnitz.de



        Abstract. Because the 2013 Cultural Heritage in CLEF (CHiC) lab fo-
        cused on multilingual retrieval, our goals were the integration of Apache
        Solr in our Xtrieval framework and the evaluation of different stemmers
        available for most of the relevant languages. As there were thirteen lan-
        guages to cover, we tried to find a generic stemmer which works with
        all languages. We experimented with four setups, where one setup was
        without any stemmer, two setups used mainly rule-based stemmers and
        the last setup used a dictionary-based stemmer. For the dictionary-based
        stemmer we employed the HunSpell stemmer, which works with the same
        dictionaries as OpenOffice.

        Keywords: stemmer, evaluation, dictionary-based stemmer, rule-based
        stemmer, cultural heritage


1     Introduction
In 2013 the Cultural Heritage in CLEF (CHiC) lab focused on multilingual
retrieval, i.e. searching over all available languages. For this task topics and
relevance assessments were provided for a total of thirteen languages. This set
allowed us to perform a large scale evaluation of stemming algorithms for all
available languages.
    Our goals for this year were the integration of Apache Solr1 in our Xtrieval
framework [1] and the evaluation of the different stemmers available within Solr
for a broad use on multilingual corpora. Hence we focused on the multilanguage
ad-hoc task.


2     System overview and setup
As usual we used our Xtrieval framework to carry out all our experiments. But
this year we added a new retrieval engine: Apache Solr. It is a very popular
open source enterprise search, [2] which is built on top of Apache Lucene2 , and
provides several web service interfaces to conduct the different tasks necessary
to perform large scale searches.
1
    http://lucene.apache.org/solr
2
    http://lucene.apache.org/java
    Although we have already been using Apache Lucene for our experiments
since 2006 [3], this new approach has changed the way we have to configure
them. A considerable amount of configuration is now done in Solr using various
XML files. Within these configuration files one can define fields with associated
types where the processing steps are declared.


                                   Xtrieval framework


                                                                    CHiC ad-hoc
    CHiC                   no stemmer                               topic
    documents
                           standard stemmer

                           less aggressive stemmer
                                                                    results in
                           HunSpell stemmer                         TREC format

                                    + stopword filter

                                        Apache Solr



    Fig. 1. Xtrieval framework using Apache Solr for our stemming experiments


    As shown in figure 1, the Xtrieval framework was still an essential part of our
experiments. We used it to read and parse the document collection and fed it
into the Solr web service. At this point there are still some pre-processing steps
left to the Xtrieval framework, which cannot be configured in Solr.
    After the indexing, we used our framework to transform the topics into search
queries for the Solr web service and gathered the results in the TREC format in
order to submit them.
    This year we focused on the different stemming approaches, which are offered
by Solr. As there were thirteen languages available, we tried to find a generic
stemmer which covers all these languages. We experimented with four settings:
one was without any stemmer, two setups used mainly rule-based stemmers
and the last setup used a dictionary-based stemmer. For the dictionary-based
stemmer we employed the HunSpell stemmer [4], which works with the same
dictionaries as OpenOffice and LibreOffice. Because of the open source nature
of these applications there are large numbers of dictionaries available for almost
every language.
    Finally, we set up these stemming experiments:
 – No stemmer
 – The standard stemmer configured in Solr: In most cases this is the snowball
   stemmer or another rule-based stemmer. Where no other stemming algo-
   rithm was available the HunSpell stemmer was used.
 – A less aggressive stemmer: This was used instead of the standard stemmer
   for every language where a less aggressive stemmer was available.
 – The HunSpell stemmer: Only languages with HunSpell dictionaries which
   had a reasonable performance used this stemmer. There were performance
    problems with complex languages, like for example French, where probably
    many connections exist between dictionary entries. The bad performance
    for this language sometimes resulted in a processing time 1000 times longer
    than with the standard stemmer. Therefore we used no stemmer for these
    languages.

   Furthermore we applied a language detection to determine fields with wrongly
labeled languages and assigned the detected language. We also removed stop
words for each language. Each language was indexed into the same index, but
using separate fields. When processing the topics for each available language a
query was issued to the whole index, i.e. every language in the index.
   The mapping of the XML data to the fields of the Solr/Lucene index was the
same we used for our experiments last year. [5]


3   Results
Table 1 shows the results of our experiments. We compared four different mea-
sures: mean average precision (map), geometric mean average precision (gmap),
binary preference (bpref), and precision at R (r-precision). Each line represents
a stemmer configuration as described in the system overview and setup. For each
measure the best score is highlighted.


                  Table 1. Results of our stemming experiments

             stemmer            map     gmap     bpref    r-precision
             standard          0.2336   0.1423   0.3115        0.3130
             less aggressive   0.2338   0.1421   0.3108        0.3140
             hunspell          0.1739   0.0928   0.2468        0.2517
             no stemmer        0.1534   0.0575   0.2239        0.2117



    The difference between standard and less aggressive stemming is marginal
and cannot be rated as significant. Without any stemmer, the results are the
lowest and the HunSpell stemmer performed slightly better than no stemming
at all.


4   Conclusions and future work
It is evident that stemming improves the results, as the experiment with no
stemming scored below every other stemming approach. Despite HunSpell stem-
ming scoring higher than no stemming, it should not be considered a beneficial
approach, because the results do not correspond to the processing time that
must be dedicated to the stemming. However, there are languages without any
other stemming algorithms, which could benefit from the HunSpell stemming.
This should be investigated further, especially in matters of processing time for
these languages.
    Another benefit of our participation was the addition of the Apache Solr
interface. Now the Xtrieval framework is able to use Solr for indexing and re-
trieving documents. Furthermore existing retrieval interfaces like AJAX Solr3
can be used to inspect the index and the documents in a more interactive way
than before.


References
[1] Kürsten, J., Wilhelm, T.: Extensible retrieval and evaluation framework: Xtrieval.
    In Baumeister, J., Atzmüller, M., eds.: LWA. Volume 448 of Technical Report.,
    Department of Computer Science, University of Würzburg, Germany (2008) 107–
    110
[2] Smiley, D., Pugh, E.: Apache solr 3 enterprise search server (2011)
[3] Kürsten, J., Eibl, M.: Monolingual retrieval experiments with a domain-specific
    document corpus at the chemnitz university of technology. [6] 178–185
[4] Halácsy, P., Trón, V.: Benefits of resource-based stemming in hungarian information
    retrieval. [6] 99–106
[5] Kürsten, J., Wilhelm, T., Richter, D., Eibl, M.: Chemnitz at the chic evaluation
    lab 2012: Creating an xtrieval module for semantic enrichment. In Forner, P., Karl-
    gren, J., Womser-Hacker, C., eds.: CLEF (Online Working Notes/Labs/Workshop).
    (2012)
[6] Peters, C., Clough, P., Gey, F.C., Karlgren, J., Magnini, B., Oard, D.W., de Rijke,
    M., Stempfhuber, M., eds.: Evaluation of Multilingual and Multi-modal Infor-
    mation Retrieval, 7th Workshop of the Cross-Language Evaluation Forum, CLEF
    2006, Alicante, Spain, September 20-22, 2006, Revised Selected Papers. In Peters,
    C., Clough, P., Gey, F.C., Karlgren, J., Magnini, B., Oard, D.W., de Rijke, M.,
    Stempfhuber, M., eds.: CLEF. Volume 4730 of Lecture Notes in Computer Science.,
    Springer (2007)




3
    https://github.com/evolvingweb/ajax-solr/