<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Overview of Morpho Challenge in CLEF 2007</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Mikko</forename><surname>Kurimo</surname></persName>
							<email>mikko.kurimo@tkk.fi</email>
							<affiliation key="aff0">
								<orgName type="department">Adaptive Informatics Research Centre</orgName>
								<orgName type="institution">Helsinki University of Technology</orgName>
								<address>
									<postBox>P.O.Box 5400</postBox>
									<postCode>FIN-02015</postCode>
									<settlement>TKK</settlement>
									<country key="FI">Finland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Mathias</forename><surname>Creutz</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Adaptive Informatics Research Centre</orgName>
								<orgName type="institution">Helsinki University of Technology</orgName>
								<address>
									<postBox>P.O.Box 5400</postBox>
									<postCode>FIN-02015</postCode>
									<settlement>TKK</settlement>
									<country key="FI">Finland</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Ville</forename><surname>Turunen</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Adaptive Informatics Research Centre</orgName>
								<orgName type="institution">Helsinki University of Technology</orgName>
								<address>
									<postBox>P.O.Box 5400</postBox>
									<postCode>FIN-02015</postCode>
									<settlement>TKK</settlement>
									<country key="FI">Finland</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Overview of Morpho Challenge in CLEF 2007</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">5187A10D5A72B1F62A3A729CDE105CC8</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-25T07:07+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>H.3 [Information Storage and Retrieval]: H.3.1 Content Analysis and Indexing; H.3.3 Information Search and Retrieval Algorithms</term>
					<term>Performance</term>
					<term>Experimentation Morphological analysis</term>
					<term>Machine learning</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Morpho Challenge 2007 contained an evaluation of unsupervised morpheme analysis algorithms using information retrieval experiments utilizing data available in CLEF. The objective of the challenge was to design statistical machine learning algorithms that discover which morphemes (smallest individually meaningful units of language) words consist of. Ideally, these are basic vocabulary units suitable for different tasks, such as text understanding, machine translation, information retrieval, and statistical language modeling The evaluation of the submitted morpheme analysis was performed by two complementary ways: Competition 1: The proposed morpheme analyses were compared to a linguistic morpheme analysis gold standard by matching the morphemesharing word pairs. Competition 2: Information retrieval (IR) experiments were performed, where the words in the documents and queries were replaced by their proposed morpheme representations and the search was based on morphemes instead of words. This paper provides an overview of the IR evaluation. The IR evaluations were provided for Finnish, German, and English and participants were encouraged to apply their algorithm to all of them. The organizers performed the IR experiments using the queries, texts, and relevance judgments available in CLEF forum and morpheme analysis methods submitted by the challenge participants. The results show that the morpheme analysis has a significant effect in IR performance in all languages, and that the performance of the best unsupervised methods can be superior to the supervised reference methods. The challenge was part of the EU Network of Excellence PASCAL Challenge Program and organized in collaboration with CLEF.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>The scientific objectives of the Morpho Challenge 2007 were: to learn of the phenomena underlying word construction in natural languages, to advance machine learning methodology, and to discover approaches suitable for a wide range of languages. The suitability for a wide range of languages is becoming increasingly important, because language technology methods need to be quickly and as automatically as possible extended to new languages that have limited previous resources. That is why learning the morpheme analysis directly from large text corpora using unsupervised machine learning algorithms is such an attractive approach and a very relevant research topic today.</p><p>The problem of learning the morphemes directly from large text corpora using an unsupervised machine learning algorithm is clearly a difficult one. First the words should be somehow segmented into meaningful parts, and then these parts should be clustered in the abstract classes of morphemes that would be useful for modeling. It is also challenging to learn to generalize the analysis to rare words, because even the largest text corpora are very sparse, a significant portion of the words may occur only once. Many important words, for example proper names and their inflections or some forms of long compound words, may also not exist in the training material at all, and their analysis is often even more challenging. However, benefits for successful morpheme analysis, in addition to obtaining a set of basic vocabulary units for modeling, can be seen for many important tasks in language technology. The additional information included in the units can provide support for building more sophisticated language models, for example, in speech recognition <ref type="bibr" target="#b0">[1]</ref>, machine translation <ref type="bibr" target="#b9">[10]</ref>, and information retrieval <ref type="bibr" target="#b12">[13]</ref>.</p><p>The evaluation of the unsupervised morpheme analysis was in this challenge solved by developing two complementary evaluations, one including a comparison to linguistic morpheme analysis gold standard, and another including a practical real-world application where morpheme analysis might be useful. This paper presents an overview how the application-oriented evaluation called Competition 2 was performed in the domain of finding useful index terms for information retrieval tasks in multiple languages using the queries, texts, and relevance judgments available in CLEF forum and morpheme analysis methods submitted by the challenge participants. The linguistic evaluation called Competition 1 are described in <ref type="bibr" target="#b7">[8]</ref> and Competition 2 in more detail in <ref type="bibr" target="#b6">[7]</ref>.</p><p>Traditionally, and especially in processing English texts, stemming algorithms have been used to reduce the different infected word forms into the common roots or stems for indexing. However, to achieve best results when ported to new languages the development of stemming algorithms requires a considerable amount of special development work. In many highly-inflecting, compounding, and agglutinative European languages the amount of different word forms is huge and the task of extracting the useful index terms becomes both more complex and more important.</p><p>The same IR tasks that were attempted using the Morpho Challenge participants' morpheme analysis, were also tested by a number of reference methods to see how the unsupervised morpheme analysis performed in comparison to them. These references included the organizers' public Morfessor Categories-Map <ref type="bibr" target="#b2">[3]</ref> and Morfessor Baseline <ref type="bibr" target="#b1">[2,</ref><ref type="bibr" target="#b3">4]</ref>, the Morfessor analysis improved by a hybrid method <ref type="bibr" target="#b11">[12]</ref>, grammatical morpheme analysis based on the linguistic gold standards <ref type="bibr" target="#b4">[5]</ref>, the traditional Porter stemming <ref type="bibr" target="#b10">[11]</ref> of words and also by the words as such without any processing.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Task</head><p>Morpho Challenge 2007 is a follow-up to our previous Morpho Challenge 2005 (Unsupervised Segmentation of Words into Morphemes) <ref type="bibr" target="#b8">[9]</ref>. In Morpho Challenge 2005 the focus was in the segmentation of data into units that are useful for statistical modeling. The specific task for the competition was to design an unsupervised statistical machine learning algorithm that segments words into the smallest meaning-bearing units of language, morphemes. In addition to comparing the obtained morphemes to a linguistic "gold standard", their usefulness was evaluated by using them for training statistical language models for speech recognition.</p><p>In Morpho Challenge 2007 a more general focus was chosen to not only to segment words into smaller units, but also to perform morpheme analysis of the word forms in the data. For instance, the English words "boot, boots, foot, feet" might obtain the analyses "boot, boot + plural, foot, foot + plural", respectively. In linguistics, the concept of morpheme does not necessarily directly correspond to a particular word segment but to an abstract class. For some languages there exist carefully constructed linguistic tools for this kind of analysis, although not for many, but using statistical machine learning methods we may still discover interesting alternatives that may rival even the most careful linguistically designed morphologies.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Training data</head><p>The Morpho Challenge 2007 task, in practice, was to return the unsupervised morpheme analysis of every word form contained in a long word list supplied by the organizers for each test language <ref type="bibr" target="#b7">[8]</ref>. The participants were pointed to corpora <ref type="bibr" target="#b7">[8]</ref> in which the words occur, so that the algorithms may utilize information about word context. The text corpora where the word list were collected were obtained from the Wortschatz collection<ref type="foot" target="#foot_0">1</ref> . at the University of Leipzig (Germany). We used the plain text files (sentences.txt for each language); the corpus sizes are 3 million sentences for English, Finnish and German, and 1 million sentences for Turkish. For English, Finnish and Turkish we used preliminary corpora, which have not yet been released publicly at the Wortschatz site. The corpora were specially preprocessed for the Morpho Challenge (tokenized, lower-cased, some conversion of character encodings).</p><p>To achieve the goal of designing language independent methods, the participants were encouraged to submit results in all test languages. The information retrieval (IR) experiments were performed by the organizers based on the morpheme analyses submitted by the participants.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">IR evaluation data</head><p>The data sets for testing the IR performance in each test language consisted of news paper articles as the source documents, test queries and the binary relevance judgments regarding to the queries. The organizers performed the IR experiments based on the morpheme analyses submitted by the participants, so it was not necessary for the participants to get these data sets. However, all the data was available for registered participants in the Cross-Language Evaluation Forum (CLEF) <ref type="foot" target="#foot_1">2</ref> .</p><p>The source documents were news articles collected from different newspapers selected as follows:</p><p>• In Finnish: 55K documents from short articles in Aamulehti 1994-95, 50 test queries on specific news topics and 23K binary relevance assessments (CLEF 2004)</p><p>• In English: 170K documents from short articles in Los Angeles Times 1994 and Glasgow Herald 1995, 50 test queries on specific news topics and 20K binary relevance assessments (CLEF 2005).</p><p>• In German: 300K documents from short articles in Frankfurter Rundschau 1994, Der Spiegel 1994-95 and SDA German 1994-95, 60 test queries with 23K binary relevance assessments (CLEF 2003).</p><p>When performing the indexing and retrieval experiments for Competition 2, it turned out that the test data contained quite many new words in addition to those that were provided as training data for the Competition 1 <ref type="bibr" target="#b7">[8]</ref>. Thus, the participants were offered a chance to improve the retrieval results of their morpheme analyses by providing them a list of the new words found in all test languages. The participants then had the choice to either run their algorithms to analyze as many of the new words as they could or liked, or to provide no extra analyses. No text data resources to find context for the new words were provided, but it was made possible to register to CLEF to use the text data available in there or any other data the participants could get.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Participants and their submissions</head><p>By the deadline in May, 2007, 6 research groups had submitted the segmentation results obtained by their algorithms. A total of 12 different algorithms were submitted, 8 of them ran experiments The organizers Helsinki Univ. Tech, FI "Morfessor Baseline"</p><p>The organizers Helsinki Univ. Tech, FI "dummy"</p><p>The organizers Helsinki Univ. Tech, FI "grammatical"</p><p>The organizers Helsinki Univ. Tech, FI "Porter"</p><p>The organizers Helsinki Univ. Tech, FI "Tepper" Michael Tepper Univ. Washington, USA on all four test languages. All the submitted algorithms are listed in Table <ref type="table" target="#tab_0">1</ref>. In addition to the competitors' 12 morpheme analysis algorithms, we evaluated a number of reference methods:</p><p>1. Public baseline methods called "Morfessor Baseline" and "Morfessor Categories-MAP" (or here just "Morfessor MAP") developed by the organizers <ref type="bibr" target="#b2">[3]</ref>.</p><p>2. No words were split nor any morpheme analysis provided, "dummy".</p><p>3. The words were analyzed using the gold standard in each language that were utilized as the "ground truth" in the Competition 1 <ref type="bibr" target="#b7">[8]</ref>. Besides the stems and suffixes, the gold standard analyses typically consist of all kinds of grammatical tags which we decided to simply include as index terms, as well. "grammatical first" uses only the first interpretation of each word whereas "grammatical all" use all.</p><p>4. Porter: No real morpheme analysis was performed, but the words were stemmed by the Porter stemming, an option provided by the Lemur toolkit.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Tepper:</head><p>A hybrid method developed by Michael Tepper <ref type="bibr" target="#b11">[12]</ref> was utilized to improve the morpheme analysis reference obtained by our Morfessor Categories-MAP.</p><p>The outputs of the submitted algorithms are analyzed closer in <ref type="bibr" target="#b7">[8]</ref>. From the IR point of view it is interesting to note that only Monson and Zeman decided to provide several alternative analysis for most words instead of just the most likely one. McNamee's algorithms did not attempt to provide a real morpheme analysis, but focused directly on finding a representative substring for each word type that would be likely to perform well in the IR evaluation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Evaluation</head><p>In this evaluation, the organizers applied the analyses provided by the participants in information retrieval experiments. The words in the queries and source documents were replaced by the corresponding morpheme analyses provided by the participants, and the search was then based on morphemes instead of words.</p><p>The evaluation was performed using a state-of-the-art retrieval method (the latest version of the freely available LEMUR toolkit<ref type="foot" target="#foot_2">3</ref> . We utilized two standard retrieval method: Tfidf and Okapi term weighting. The Tfidf implementation in LEMUR applies term frequency weights for both query and document based on the BM25 weighting and the Euclidean dot-product as similarity measure. Okapi in LEMUR is an implementation of the BM25 retrieval function as described in <ref type="bibr" target="#b5">[6]</ref>.</p><p>The evaluation criterion was Uninterpolated Average Precision There were several different categories and the winner with the highest Average Precision was selected separately for each language and each category:</p><p>1. All morpheme analyses from the training data are used as index terms "withoutnew" vs.</p><p>additionally using also the morpheme analyses for new words that existed in the IR data but not in the training data "withnew".</p><p>2. Tfidf term weighting was utilized for all index terms without any stoplists vs. Okapi term weighting for all index terms excluding an automatic stoplist consisting of the most common terms (frequency threshold was 75,000 for Finnish and 150,000 for German and English).</p><p>The stoplist was developed for the Okapi weighting, because otherwise Okapi weights were not suitable for the indexes that had many very common terms.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">Results</head><p>The results of the information retrieval evaluations are shown in Table <ref type="table" target="#tab_1">2</ref>. Here we have selected only the best runs from each participant (in bold) and reference method. For the full results see <ref type="bibr" target="#b6">[7]</ref>. Indexing is performed using Tfidf weighting for all morphemes (left) and Okapi weighting for all morphemes except the most common ones (stoplist) with frequency higher than 150,000 (right).</p><p>In the Finnish task, the highest average precision was obtained by the "Bernhard 2" algorithm, which was also won the Competition 1 <ref type="bibr" target="#b7">[8]</ref>. The highest average precision 0.49 was obtained using the Okapi weighting and stoplist for both the originally submitted morpheme analysis (for Competition 1) and the morpheme analysis for the new words added for Competition 2. The "Bernhard 1" algorithm obtained the highest average precision 0.47 for the German task using the new words, Okapi and stoplist. For English, the highest average precision was obtained by the "Bernhard 2" algorithm, which was also won the Competition 1 <ref type="bibr" target="#b7">[8]</ref>. As in Finnish and German, the highest average precision 0.39 was obtained with the new words and using the Okapi weighting and stoplist.</p><p>As expected, the "grammatical" reference method based on linguistic Gold Standard morpheme analysis <ref type="bibr" target="#b7">[8]</ref> did not perform very well. However, with stoplist and Okapi term weighting it did achieve better results than the "dummy" method in all languages. In Finnish and English the performance was better than average, but quite poor in German. The "grammatical first" that utilized only the first of the alternative analysis in indexing was at least as good or better than the "grammatical all", which seems to indicate that the alternative analysis are not very useful here.</p><p>For the "Morfessor" references it is interesting to note that they always performed better than the "grammatical", which seems to suggest that the coverage of the analysis ("Morfessor" does not have any out-of-vocabulary words) is more important for IR than the grammatical correctness. In general, the old "Morfessor Baseline" seems to provide a very good baseline in all tested languages also for the IR tasks as it did for the language modeling and speech recognition in <ref type="bibr" target="#b8">[9]</ref>. The comparison of the results in the Tfidf and Okapi categories show that the Okapi with stoplist performed significantly better for all languages. We also run Tfidf with stoplist (the results not included here) which achieved results that were better than the plain Tfidf and only slightly inferior to Okapi with stoplist. However, we decided to rather report the original Tfidf, since we wanted to show what is the performance and the relative ranking of the methods without the stoplist. The Porter stemming that is a standard word preprocessing tool in IR remained unbeaten (by a narrow margin) in our evaluations in English, but in German and especially in Finnish, the unsupervised morpheme analysis methods clearly dominated the evaluation. There might exist better stemming algorithms for those languages, but because of the more complex morphology, their development might not be an easy task.</p><p>As future work in this field it should be relatively straight-forward to evaluate the unsupervised morpheme analysis in several other interesting languages, because it is not bounded to only those languages where rule-based grammatical analysis can be performed. It would also be interesting to try to combine the rival analysis to produce something better.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="9">Conclusions</head><p>The objective of Morpho Challenge 2007 was to design a statistical machine learning algorithm that discovers which morphemes (smallest individually meaningful units of language) words consist of. Ideally, these are basic vocabulary units suitable for different tasks, such as text understanding, machine translation, information retrieval, and statistical language modeling. The current challenge was a successful follow-up to our previous Morpho Challenge 2005 (Unsupervised Segmentation of Words into Morphemes). This time the task was more general in that instead of looking for an explicit segmentation of words, the focus was in the morpheme analysis of the word forms in the data.</p><p>The scientific goals of this challenge were to learn of the phenomena underlying word construction in natural languages, to discover approaches suitable for a wide range of languages and to advance machine learning methodology. The analysis and evaluation of the submitted machine learning algorithm for unsupervised morpheme analysis showed that these goals were quite nicely met. There were several novel unsupervised methods that achieved good results in several test languages, both with respect to finding meaningful morphemes and useful units for information retrieval. The IR results also revealed that the morpheme analysis has a significant effect in IR performance in all languages, and that the performance of the best unsupervised methods can be superior to the supervised reference methods.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>The submitted algorithms.</figDesc><table><row><cell>Algorithm</cell><cell>Authors</cell><cell>Affiliation</cell></row><row><cell>"Bernhard 1"</cell><cell>Delphine Bernhard</cell><cell>TIMC-IMAG, F</cell></row><row><cell>"Bernhard 2"</cell><cell>Delphine Bernhard</cell><cell>TIMC-IMAG, F</cell></row><row><cell>"Bordag 5"</cell><cell>Stefan Bordag</cell><cell>Univ. Leipzig, D</cell></row><row><cell>"Bordag 5a"</cell><cell>Stefan Bordag</cell><cell>Univ. Leipzig, D</cell></row><row><cell>"McNamee 3"</cell><cell cols="2">Paul McNamee and James Mayfield JHU, USA</cell></row><row><cell>"McNamee 4"</cell><cell cols="2">Paul McNamee and James Mayfield JHU, USA</cell></row><row><cell>"McNamee 5"</cell><cell cols="2">Paul McNamee and James Mayfield JHU, USA</cell></row><row><cell>"Zeman "</cell><cell>Daniel Zeman</cell><cell>Karlova Univ., CZ</cell></row><row><cell>"Monson Morfessor"</cell><cell>Christian Monson et al.</cell><cell>CMU, USA</cell></row><row><cell>"Monson ParaMor"</cell><cell>Christian Monson et al.</cell><cell>CMU, USA</cell></row><row><cell cols="2">"Monson ParaMor-Morfessor" Christian Monson et al.</cell><cell>CMU, USA</cell></row><row><cell>"Pitler"</cell><cell>Emily Pitler and Samarth Keshava</cell><cell>Univ. Yale, USA</cell></row><row><cell>"Morfessor Categories-MAP"</cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2 :</head><label>2</label><figDesc>The obtained average precision (AP%) in the information retrieval task for the best submitted segmentation for each participant and reference method.</figDesc><table><row><cell>Finnish:</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="2">Tfidf weighting for all morphemes</cell><cell></cell><cell cols="2">Okapi weighting and a stoplist</cell></row><row><cell>METHOD</cell><cell cols="2">WORDLIST AP%</cell><cell>METHOD</cell><cell cols="2">WORDLIST AP%</cell></row><row><cell>Morfessor baseline</cell><cell>withnew</cell><cell cols="2">0.4105 Bernhard 2</cell><cell>withnew</cell><cell>0.4915</cell></row><row><cell>Bernhard 1</cell><cell>withoutnew</cell><cell cols="2">0.4016 Morfessor baseline</cell><cell>withnew</cell><cell>0.4412</cell></row><row><cell>grammatical first</cell><cell>withoutnew</cell><cell cols="2">0.3995 Bordag 5a</cell><cell>withnew</cell><cell>0.4309</cell></row><row><cell>Bordag 5</cell><cell>withnew</cell><cell cols="2">0.3831 grammatical all</cell><cell>withoutnew</cell><cell>0.4307</cell></row><row><cell>McNamee 5</cell><cell>withoutnew</cell><cell cols="2">0.3646 McNamee 5</cell><cell>withnew</cell><cell>0.3684</cell></row><row><cell>Porter</cell><cell>withnew</cell><cell cols="2">0.3566 Porter</cell><cell>withnew</cell><cell>0.3517</cell></row><row><cell>dummy</cell><cell>withnew</cell><cell cols="2">0.3559 dummy</cell><cell>withnew</cell><cell>0.3274</cell></row><row><cell>Zeman</cell><cell>withoutnew</cell><cell cols="2">0.2494 Zeman</cell><cell>withoutnew</cell><cell>0.2813</cell></row><row><cell>German:</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="2">Tfidf weighting for all morphemes</cell><cell></cell><cell cols="2">Okapi weighting and a stoplist</cell></row><row><cell>METHOD</cell><cell cols="2">WORDLIST AP%</cell><cell>METHOD</cell><cell cols="2">WORDLIST AP%</cell></row><row><cell>Morfessor baseline</cell><cell>withnew</cell><cell cols="2">0.3874 Bernhard 1</cell><cell>withnew</cell><cell>0.4729</cell></row><row><cell>Bernhard 1</cell><cell>withoutnew</cell><cell cols="3">0.3777 Monson Morfessor withnew</cell><cell>0.4602</cell></row><row><cell>Porter</cell><cell>withnew</cell><cell cols="2">0.3725 Morfessor MAP</cell><cell>withnew</cell><cell>0.4571</cell></row><row><cell cols="2">Monson Morfessor withnew</cell><cell cols="2">0.3520 Bordag 5</cell><cell>withnew</cell><cell>0.4308</cell></row><row><cell>dummy</cell><cell>withnew</cell><cell cols="2">0.3496 Porter</cell><cell>withnew</cell><cell>0.3866</cell></row><row><cell>Bordag 5a</cell><cell>withnew</cell><cell cols="2">0.3496 McNamee 5</cell><cell>withoutnew</cell><cell>0.3617</cell></row><row><cell>McNamee 5</cell><cell>withoutnew</cell><cell cols="2">0.3442 grammatical first</cell><cell>withoutnew</cell><cell>0.3467</cell></row><row><cell>grammatical first</cell><cell>withoutnew</cell><cell cols="2">0.3223 dummy</cell><cell>withnew</cell><cell>0.3228</cell></row><row><cell>Zeman</cell><cell>withoutnew</cell><cell cols="2">0.2828 Zeman</cell><cell>withoutnew</cell><cell>0.2568</cell></row><row><cell>English:</cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell cols="2">Tfidf weighting for all morphemes</cell><cell></cell><cell cols="2">Okapi weighting and a stoplist</cell></row><row><cell>METHOD</cell><cell cols="2">WORDLIST AP%</cell><cell>METHOD</cell><cell cols="2">WORDLIST AP%</cell></row><row><cell>Porter</cell><cell>withnew</cell><cell cols="2">0.3052 Porter</cell><cell>withnew</cell><cell>0.4083</cell></row><row><cell>McNamee 5</cell><cell>withoutnew</cell><cell cols="2">0.2888 Bernhard 2</cell><cell>withnew</cell><cell>0.3943</cell></row><row><cell>Morfessor baseline</cell><cell>withnew</cell><cell cols="2">0.2863 Morfessor baseline</cell><cell>withnew</cell><cell>0.3882</cell></row><row><cell>Tepper</cell><cell>withoutnew</cell><cell cols="2">0.2784 grammatical first</cell><cell>withoutnew</cell><cell>0.3774</cell></row><row><cell>dummy</cell><cell>withnew</cell><cell cols="2">0.2783 Tepper</cell><cell>withoutnew</cell><cell>0.3728</cell></row><row><cell>Bernhard 1</cell><cell>withoutnew</cell><cell cols="3">0.2781 Monson Morfessor withoutnew</cell><cell>0.3721</cell></row><row><cell cols="2">Monson Morfessor withoutnew</cell><cell cols="2">0.2676 Pitler</cell><cell>withoutnew</cell><cell>0.3652</cell></row><row><cell>Pitler</cell><cell>withoutnew</cell><cell cols="2">0.2666 McNamee 4</cell><cell>withoutnew</cell><cell>0.3577</cell></row><row><cell>grammatical all</cell><cell>withoutnew</cell><cell cols="2">0.2619 Bordag 5</cell><cell>withoutnew</cell><cell>0.3427</cell></row><row><cell>Zeman</cell><cell>withoutnew</cell><cell cols="2">0.2297 dummy</cell><cell>withnew</cell><cell>0.3123</cell></row><row><cell>Bordag 5</cell><cell>withoutnew</cell><cell cols="2">0.2210 Zeman</cell><cell>withoutnew</cell><cell>0.2674</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">http://corpora.informatik.uni-leipzig.de/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">http://www.clef-campaign.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">http://www.lemurproject.org/</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>We thank all the participants for their submissions and enthusiasm. We owe great thanks as well to the organizers of the PASCAL Challenge Program and CLEF who helped us organize this challenge and the challenge workshop. Especially, we would like to thank Carol Peters from CLEF for helping us to get Morpho Challenge in CLEF 2007 and organize a great workshop there. Our work was supported by the Academy of Finland in the projects Adaptive Informatics and New adaptive and learning methods in speech recognition. This work was supported in part by the IST Programme of the European Community, under the PASCAL Network of Excellence, IST-2002-506778. This publication only reflects the authors' views. We acknowledge that access rights to data and other materials are restricted due to other commitments.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Factored language models and generalized parallel backoff</title>
		<author>
			<persName><forename type="first">Jeff</forename><forename type="middle">A</forename><surname>Bilmes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Katrin</forename><surname>Kirchhoff</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Human Language Technology, Conference of the North American Chapter of the Association for Computational Linguistics (HLT-NAACL)</title>
				<meeting>the Human Language Technology, Conference of the North American Chapter of the Association for Computational Linguistics (HLT-NAACL)<address><addrLine>Edmonton, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2003">2003</date>
			<biblScope unit="page" from="4" to="6" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Unsupervised discovery of morphemes</title>
		<author>
			<persName><forename type="first">Mathias</forename><surname>Creutz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Krista</forename><surname>Lagus</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Workshop on Morphological and Phonological Learning of ACL-02</title>
				<meeting>the Workshop on Morphological and Phonological Learning of ACL-02</meeting>
		<imprint>
			<date type="published" when="2002">2002</date>
			<biblScope unit="page" from="21" to="30" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Inducing the morphological lexicon of a natural language from unannotated text</title>
		<author>
			<persName><forename type="first">Mathias</forename><surname>Creutz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Krista</forename><surname>Lagus</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International and Interdisciplinary Conference on Adaptive Knowledge Representation and Reasoning (AKRR&apos;05)</title>
				<meeting>the International and Interdisciplinary Conference on Adaptive Knowledge Representation and Reasoning (AKRR&apos;05)<address><addrLine>Espoo, Finland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2005">2005</date>
			<biblScope unit="page" from="106" to="113" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Unsupervised morpheme segmentation and morphology induction from text corpora using Morfessor</title>
		<author>
			<persName><forename type="first">Mathias</forename><surname>Creutz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Krista</forename><surname>Lagus</surname></persName>
		</author>
		<ptr target="http://www.cis.hut.fi/projects/morpho/" />
		<imprint>
			<date type="published" when="2005">2005</date>
		</imprint>
		<respStmt>
			<orgName>Publications in Computer and Information Science, Helsinki University of Technology</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical Report A81</note>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Morpheme segmentation gold standards for finnish and english</title>
		<author>
			<persName><forename type="first">Mathias</forename><surname>Creutz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Krister</forename><surname>Linden</surname></persName>
		</author>
		<ptr target="http://www.cis.hut.fi/projects/morpho/" />
		<imprint>
			<date type="published" when="2004">2004</date>
		</imprint>
		<respStmt>
			<orgName>Publications in Computer and Information Science, Helsinki University of Technology</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical Report A77</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Okapi at TREC-3</title>
		<author>
			<persName><forename type="first">S</forename><surname>Robertson</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Third Text Retrieval Conference (TREC-3)</title>
				<meeting>the Third Text Retrieval Conference (TREC-3)</meeting>
		<imprint>
			<date type="published" when="1994">1994</date>
			<biblScope unit="page" from="109" to="126" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Unsupervised morpheme analysis evaluation by IR experiments -Morpho Challenge</title>
		<author>
			<persName><forename type="first">Mikko</forename><surname>Kurimo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mathias</forename><surname>Creutz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ville</forename><surname>Turunen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes for the CLEF 2007 Workshop</title>
				<meeting><address><addrLine>Budapest, Hungary</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2007">2007. 2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Unsupervised morpheme analysis evaluation by a comparison to a linguistic Gold Standard -Morpho Challenge</title>
		<author>
			<persName><forename type="first">Mikko</forename><surname>Kurimo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mathias</forename><surname>Creutz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Matti</forename><surname>Varjokallio</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes for the CLEF 2007 Workshop</title>
				<meeting><address><addrLine>Budapest, Hungary</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2007">2007. 2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Unsupervised segmentation of words into morphemes -Challenge 2005, an introduction and evaluation report</title>
		<author>
			<persName><forename type="first">Mikko</forename><surname>Kurimo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mathias</forename><surname>Creutz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Matti</forename><surname>Varjokallio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ebru</forename><surname>Arisoy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Murat</forename><surname>Saraclar</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">PASCAL Challenge Workshop on Unsupervised segmentation of words into morphemes</title>
				<meeting><address><addrLine>Venice, Italy</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Morphological analysis for statistical machine translation</title>
		<author>
			<persName><forename type="first">Y.-S</forename><surname>Lee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Human Language Technology, Conference of the North American Chapter of the Association for Computational Linguistics (HLT-NAACL)</title>
				<meeting>the Human Language Technology, Conference of the North American Chapter of the Association for Computational Linguistics (HLT-NAACL)<address><addrLine>Boston, MA, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">An algorithm for suffix stripping</title>
		<author>
			<persName><forename type="first">M</forename><surname>Porter</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Program</title>
		<imprint>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="130" to="137" />
			<date type="published" when="1980-07">July 1980</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">A Hybrid Approach to the Induction of Underlying Morphology</title>
		<author>
			<persName><forename type="first">Michael</forename><surname>Tepper</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2007">2007</date>
		</imprint>
		<respStmt>
			<orgName>University of Washington</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">PhD thesis</note>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Conceptual mapping of user&apos;s queries to medical subject headings</title>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">L</forename><surname>Zieman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">L</forename><surname>Bleich</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 1997 American Medical Informatics Association (AMIA) Annual Fall Symposium</title>
				<meeting>the 1997 American Medical Informatics Association (AMIA) Annual Fall Symposium</meeting>
		<imprint>
			<date type="published" when="1997-10">October 1997</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
