<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Intrinsic word embedding model evaluation for Lithuanian language using adapted similarity and relatedness benchmark datasets</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Mindaugas</forename><surname>Petkevičius</surname></persName>
							<email>mindaugas.petkevicius@vdu.lt</email>
							<affiliation key="aff1">
								<orgName type="institution">Vytautas Magnus University</orgName>
								<address>
									<addrLine>K. Donelaičio g. 58</addrLine>
									<postCode>44248</postCode>
									<settlement>Kaunas</settlement>
									<country key="LT">Lithuania</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Daiva</forename><surname>Vitkutė-Adžgauskienė</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">Vytautas Magnus University</orgName>
								<address>
									<addrLine>K. Donelaičio g. 58</addrLine>
									<postCode>44248</postCode>
									<settlement>Kaunas</settlement>
									<country key="LT">Lithuania</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff0">
								<orgName type="department" key="dep1">26th</orgName>
								<orgName type="department" key="dep2">International Conference Information Society</orgName>
								<orgName type="institution">University Studies -IVUS</orgName>
								<address>
									<postCode>2021</postCode>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Intrinsic word embedding model evaluation for Lithuanian language using adapted similarity and relatedness benchmark datasets</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">145B19725FE207063058768957826702</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T21:40+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Word embeddings</term>
					<term>evaluation</term>
					<term>Lithuanian language</term>
					<term>word2vec</term>
					<term>fastText</term>
					<term>GloVe</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Word embeddings are real-valued word representations capable of capturing lexical semantics and trained on natural language corpora. Word embedding models have gained popularity in recent years, but the issue of selecting the most adequate word embedding evaluation methods remains open. This paper presents research on adaptation of the intrinsic similarity and relatedness task for the Lithuanian language and the evaluation of word embedding models, testing the quality of representations independently of specific natural language processing tasks. 7 different evaluation benchmarks were adapted for the Lithuanian language and 50 word embedding models were trained using fastText, GloVe, and Word2vec algorithms and evaluated on syntactic and semantic similarity tasks. The obtained results suggest that for the intrinsic similarity and relatedness task, the dimension parameter has a significant impact on the evaluation results, with larger word embedding dimension yielding better results.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>The development of natural language processing tools influenced a growing need for word embeddings as real-valued representations of words for text analytics, generated by applying distributive semantic models. While word embeddings have become one of the most widely used tools in modern natural language processing (NLP) applications, their limitations have not yet been fully explored. The problem of assessing word embedding consistency and quality is one of the most relevant questions in distributive semantics research.</p><p>The idea of word embeddings is not new, but it gained popularity after Mikolov et al. <ref type="bibr" target="#b0">[1]</ref> presented the Word2vec model in 2013. The fastText model, developed by Facebook AI Research (FAIR), introduces embeddings using subword information. The next big improvement came from Stanford with GLoVE (Global-Vectors) <ref type="bibr" target="#b1">[2]</ref>, based on word-word co-occurrence statistics in a corpus.</p><p>There are two types of word embedding evaluation: intrinsic and extrinsic. Intrinsic evaluation tests the representation quality independent of specific natural language processing (NLP) tasks, while extrinsic evaluation uses word embeddings as input features to an NLP task and measures changes in corresponding performance metrics. We focus on intrinsic evaluation methods, based on human annotated datasets, because datasets can be adapted for different languages by translating and reevaluating human annotated scores.</p><p>The method of word semantic similarity, based on correlation with human judgment of how closely words are related among themselves, was one of the first intrinsic evaluation metrics for distributional meaning representations. According to this method, words smart and intelligent should be closer in the vector space than smart and dumb, since smart and intelligent are intuitively better semantically related.</p><p>There are gold-standard benchmarks for evaluating distributive semantic models such as SimLex999 <ref type="bibr" target="#b2">[3]</ref>, MEN <ref type="bibr" target="#b3">[4]</ref>, etc., focused on semantic relatedness. These benchmarks consist of certain word pairs and their relative similarity scores. The similarity scores are defined in the interval between 0 and 10, e.g., the score for the words book and paper is 7.46. When applied, these scores are compared with word pair cosine vector similarity results for word embeddings.</p><p>The word analogy method aims to identify words based on operation prediction in a word vector space. The method tries to predict a missing word in a word pair based on a known relationship in another word pair. Thus, for a dataset a-b, c-d, the task is to identify an unknown word d based on the known relationship between words a and b. For example, given the words, a (brother), b (sister), and c (father), this method should correctly predict the value mother for word d <ref type="bibr" target="#b4">[5]</ref>. The Google analogy dataset <ref type="bibr" target="#b5">[6]</ref> and BATS <ref type="bibr" target="#b6">[7]</ref> are the most popular datasets. The Google test set has become the standard for word embedding analysis. BATS is a newer dataset that is much larger and more balanced.</p><p>The word clustering method evaluates a word embedding space by applying the word clustering approach. It is aimed at splitting a given word set into groups of words corresponding to different categories based on word vectors. For example, words dog and cat belong to one cluster, while words car and planeto another <ref type="bibr" target="#b7">[8]</ref>.</p><p>The situation with word embeddings for the Lithuanian language is influenced by its specifics. The Lithuanian language is a morphologically rich Baltic language, being considered one of the most archaic living Indo-European languages <ref type="bibr" target="#b8">[9]</ref>. It has a relatively large vocabulary, containing over 500 000 unique words <ref type="bibr" target="#b9">[10]</ref>. On the other hand, the Lithuanian language lacks textual resources due to the small size of the nation using it. Lithuanian Wikipedia, for example, has 199 567 articles, while better represented languages have over a million each <ref type="bibr" target="#b10">[11]</ref>. Several attempts were made to perform intrinsic and extrinsic evaluation of the Lithuanian language embeddings. However, so far there are no available semantic similarity benchmarks for this purpose.</p><p>The goal of this research was to adapt selected intrinsic similarity benchmarks for the Lithuanian language and to apply them for experimental evaluation of fastText, Word2vec, and GloVe embedding models with different hyperparameters.</p><p>In order to reach this goal, we perform the following tasks: related work analysis (Section 2), corpus building for embedding training (Section 3), methodology for the adaptation of evaluation benchmarks for the Lithuanian language (Section 4), experimental evaluation of different embeddings based on the derived benchmarks (Section 5), conclusions and future plans (Section 6).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Works</head><p>In recent years, there have been several critical articles on intrinsic assessment methods: some researchers address the subjectivity of human judgments, the vagueness of instructions for particular tasks, and terminology confusions <ref type="bibr" target="#b11">[12]</ref>. However, despite these flaws, these methods are widely used for embedding model evaluation for different languages.</p><p>There have been successful attempts to adapt intrinsic evaluation benchmarks to other languages. Research has shown that when monolingual vector space models were translated into German, Russian, and Italian, it became clear that their predictions did not always correlate well with human decisions made in the language used for model training <ref type="bibr" target="#b12">[13]</ref>.</p><p>Another study attempted to translate the SymLex999 benchmark into Estonian and discovered that, unlike in the original research, computational word embedding models better correlate with noun scores rather than adjective scores <ref type="bibr" target="#b13">[14]</ref>.</p><p>A few studies on the evaluation of Lithuanian word embeddings have been carried out. In the first study, word embeddings for different models and training algorithms were evaluated against a limited implementation of the Lithuanian WordNet <ref type="bibr" target="#b14">[15]</ref>, showing that the Continuous Bag of Words (CBOW) approach performed significantly better than the skip-gram approach for Word2vec word embeddings, vector dimensions having little effect in this case.</p><p>The second study compared traditional and deep learning approaches for sentiment analysis using word embeddings, finding that deep learning performed well only when applied to small datasets, and that traditional methods performed better in all other contexts <ref type="bibr" target="#b15">[16]</ref>.</p><p>The third study was conducted with Transformer models using GloVe word embeddings <ref type="bibr" target="#b16">[17]</ref>. The study concluded that multilingual transformer models can be fine-tuned to word vectors, but still perform much worse than specifically trained embeddings.</p><p>In conclusion, we see that the Lithuanian language lacks word embedding evaluation benchmarks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Corpus</head><p>Semantic intrinsic similarity benchmarks cover many different types of test domains, such as geography, languages, currency, etc. Therefore, we need to have a wide variety of data for embedding training. Research has shown that for a larger corpus we get better word embeddings <ref type="bibr" target="#b17">[18]</ref>. For this reason, it is important to build an extensive corpus for embedding training, that will be further used for evaluation.</p><p>Wikipedia texts are usually a typical approach for building a corpus for embedding training. In order to expand our experimental corpus, we used articles from Lithuanian news portals, mainly from the largest one, Delfi.lt, the collected articles covering different topical areas such as news, cars, fitness, culture, food, and so on.</p><p>In order to obtain better word embeddings, we also included texts from the Corpus of Contemporary Lithuanian Language (CCLL) <ref type="bibr" target="#b18">[19]</ref>, texts in a variety of genres and topics.</p><p>Statistics for our combined experimental corpus is presented in Table <ref type="table" target="#tab_0">1</ref>. The pre-processing phase consists of two steps: 1) breaking text into tokens, lowercasing text, removing special symbols, numbers, non-Lithuanian words and stop-words 2) removing short documents, less than 50 characters in size; 3) lemmatizing the texts, this being important for rich morphology languages <ref type="bibr" target="#b19">[20]</ref>. Lemmatization was performed using lexical and morphological analysis tools from the Lithuanian language technology infrastructure built in the Semantika2 <ref type="bibr" target="#b20">[21]</ref> project.</p><p>Alternatively, text could have been stemmed instead, but lemmatization was preferred, as all our documents were in normative spelling and punctuation. Stemming is more favorable in case of social texts with lots of out-of-dictionary words. Also, stemming has its limitations, e.g. over-stemming and under-stemming problems <ref type="bibr" target="#b21">[22]</ref>.</p><p>The statistics for the final version of our experimental corpus are presented in Table <ref type="table" target="#tab_1">2</ref>. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Methodology</head><p>The methodology part covers the methods applied in this research: (1) benchmark dataset adaptation method for semantic similarity based intrinsic embedding model evaluation; (2) semantic similarity based embedding evaluation using the adapted benchmarks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Adaptation of benchmark datasets</head><p>As a result of a brief analysis, the following English-language benchmarks were selected for the adaptation to the Lithuanian language, their popularity being the main criteria:</p><p>1. MEN (Marco, Elia and Nam), 3 000 pairs <ref type="bibr" target="#b22">[23]</ref>. The following algorithm was applied for the dataset adaptation to the Lithuanian language: 1. Automated translation of datasets (by applying the Google Cloud Translation API) <ref type="bibr" target="#b29">[30]</ref>. 2. Inconsistency checking (manual examination), discarding inconsistent word pairs. 3. Word lemmatization. 4. Re-evalution of the score that was initially assigned to the English language word pairs was done by two independent persons (manual procedure). An average score was calculated.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Embedding evaluation using semantic similarity benchmarks</head><p>As mentioned in Chapter 1, the semantic similarity datasets are based on correlation with human judgments of how closely words are related.</p><p>The similarity benchmark datasets consist of a certain number of word pairs. Each pair is determined by its similarity and relatedness. The values are in the range [0, 10], depending on the dataset.</p><p>The word embedding models are represented by corresponding vectors for each word in the dictionary. If a word is missing in the trained word embedding model, it is replaced by the mean of all vectors. In order to calculate the similarity between vectors, we can use the cosine similarity formula (see Eq. 1), where a and b are vectors in the word embedding vector space.</p><formula xml:id="formula_0">𝑠𝑖𝑚 𝑐𝑜𝑠 (𝑎 𝑖 , 𝑏 𝑖 ) = 𝒂 𝒊 . 𝒃 𝒊 ||𝒂 𝒊 || × ||𝒃 𝒊 || ,<label>(1)</label></formula><p>here ai and bi are vectors of N-dimension. The result of cosine similarity is a value in the range [-1, 1] interval, where 1 stands for identical vectors, and -1 for opposite vectors.</p><p>A human-annotated benchmark dataset consists of n triplets containing pairs of words and their corresponding similarity scores ⟨wi , wj , hij⟩, where wi , wj are dictionary words, and hij is the score.</p><p>Let h = (hi1, hi2, . . . , hiN ) be a vector of human annotated benchmark datasets, and m = (mi1, mi2, . . . , miN ), correspondingly, a vector of similarity scores calculated from word embeddings.</p><p>Then, the evaluation score for the corresponding embedding model, based on the selected benchmark, is calculated as Spearman's correlation ρ (see Eq. 2) between h and m.</p><p>Spearman 𝑝 value can be any value satisfying−1 ≤ 𝑝 ≤ 1, and the interpretation is that p values close to +1 indicate stronger relationship, while those closer to -1 indicate weaker relationship.</p><p>The Spearman correlation formula is:</p><formula xml:id="formula_1">𝑝 = 1 − 6 ∑ 𝑑 𝑖 2 𝑛(𝑛 2 − 1) ,<label>(2)</label></formula><p>where ndataset length, ddifference between ranks of h and m.</p><p>The aggregated score of one-word embedding model 𝑝 𝑎𝑣𝑔 is calculated (see Eq. 3) as:</p><formula xml:id="formula_2">𝑝 𝑎𝑣𝑔 = 1 𝑛 * ∑ 𝑝 𝑖 𝑛 𝑖=1<label>(3)</label></formula><p>where 𝑝 𝑖 is Spearman correlation value of specific benchmark, nnumber of benchmarks.</p><p>In order to compare different embedding model types (Word2vec, fastText, GloVe). We can calculate an average score of all models' embeddings (see <ref type="bibr">Eq. 4)</ref>.</p><formula xml:id="formula_3">𝑃 𝑡 = 1 𝑛 * ∑ 𝑝 𝑎𝑣𝑔 𝑛 𝑖=1<label>(4)</label></formula><p>where 𝑃 𝑡an average score of all t word embeddings, nnumber of t word embedding models.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Experiments and results</head><p>Experiments were carried out in a series of tasks: 1. Firstly, 7 selected evaluation benchmark datasets were adapted for the Lithuanian language. 2. Secondly, 50 word embeddings with different hyperparameter sets were trained on the accumulated experimental corpus.</p><p>3. Thirdly, the obtained word embedding models were evaluated using the adapted intrinsic evaluation benchmarks. 4. Finally, the resulting data were examined in order to determine the effect of different hyperparameters on benchmark evaluation results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Adaptation of benchmark datasets</head><p>The selected 7 (see Chapter 4.1) evaluation benchmark datasets were adapted from English to Lithuanian language. There were 5610 word pairs at the beginning. After the adaptation process, 5573 word pairs remained. A total of 37 word pairs were discarded. The following problems were observed during the adaptation process:</p><p>1. Multiple wordsin some cases, one-to-one word translation is not possible, when a two-word expression in the Lithuanian language is a correspondence to a single word in the English language. For example, for the word pair "computer -software", the Lithuanian translation would be "kompiuteris -programinė įranga". As we use vector-to-vector comparison, such word pairs were discarded.</p><p>2. The meaning of certain words has been shaped by American culture, e.g. words like soccer, football, and FBI. These are words that are commonly used in the US. Such words were replaced with Lithuanian synonyms 3. A few older words have undergone semantic changes as their meanings evolved. For example, the word pair "Arafat -terror", had a greater similarity back in history than it does now. Such pairs were discarded. 4. In some cases, both English words have the same meaning in the Lithuanian language, for example, the following pairs: "smart -intelligent", "happy -cheerful", "fast -rapid". Such word pairs as a result contained two equal words, and their scores were set to 10 (maximum similarity). An excerpt of the adapted SimLex999 dataset for the Lithuanian language is presented in Table <ref type="table" target="#tab_3">3</ref>. The first two columns contain an English word pair in its original form. The third column contains the human-generated similarity score. The fourth and fifth columns contain Lithuanian translations of English words and revalued Lithuanian word scores.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2.">Word embedding model training</head><p>The following tools were used for word embedding training: python genism wrapper of Word2vec<ref type="foot" target="#foot_0">1</ref> , fastTextofficial python library<ref type="foot" target="#foot_1">2</ref> , GloVe -official library <ref type="foot" target="#foot_2">3</ref> . We used similar training parameters in order to be able to compare different word embeddings (see Table <ref type="table" target="#tab_4">4</ref>). </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3.">Word embedding model evaluation</head><p>All the trained embedding models were evaluated using the Spearman ρ correlation coefficient between human benchmark scores and vector space model scores. The results were grouped by different vector model types, characterized by different hyperparameter sets (see Table <ref type="table" target="#tab_4">4</ref>).</p><p>The best 4 and the worst 4 models ranked by the benchmark result average are presented correspondingly in Table <ref type="table" target="#tab_6">5 and Table 6</ref>  To be able to do score comparison, only embedding models with the same hyperparameters were used: dimensions (100, 300), window size (5), and minimum count (2, 5) (Figure <ref type="figure" target="#fig_1">1</ref>). GloVe's word embedding model scores 𝑃 𝑡 were on average lower than those of fastText and Word2vec. The previous two were nearly identical, with a difference of only 0,001 between them.</p><p>Additionally, the experiment results were analyzed to determine whether a particular hyperparameter had a significant effect on the results. Following a thorough examination of all the hyperparameters, we discovered a correlation between the dimension value and the correlation results. (Figure <ref type="figure" target="#fig_2">2</ref>). Different dimension values for various embedding types had a significant effect on the results. The larger the dimension of the word embedding, the more accurate the results. As illustrated in Figure <ref type="figure" target="#fig_3">3</ref>, as vector size increases, the model correlation score also increases. (5)</p><p>where nnumber of models, xdimension value. y -Spearman correlation value for a model. 𝑟 = 0.918, this indicates strong relationship between values.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusions</head><p>This was the first attempt to adapt the most popular intrinsic similarity and relatedness benchmark datasets for the Lithuanian language. Despite reported challenges when adapting benchmarks to other languages, we proved, that this can be done even for morphology rich languages like Lithuanian.</p><p>The application of the adapted benchmark datasets for the evaluation of the embedding models, trained on an experimental corpus, showed, that GloVe model performed worse than fastText and Word2vec, judging by average benchmark results.</p><p>We also conclude, that for the intrinsic similarity and relatedness task, the dimension hyperparameter has a significant impact on the evaluation results, with larger word embedding dimension yielding better results.</p><p>In the future, we plan to adapt other types of embedding evaluation benchmarks, such as categorization and analogy testing, as well as extrinsic evaluation with POS tagging, named entity recognition (NER), and other NLP tasks. This would allow us to compare intrinsic and extrinsic evaluation methods. Also, we will continue to expand our corpus for future tests.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head></head><label></label><figDesc>. The first column in these tables indicates model name together with hyperparameter indication. The following labels are used: Nnegative sampling, S -SoftMax, CBOW -Continuous Bag of Words, SKIP -Skipgram, ddimension, wwindow size, mminimum count threshold, iiteration count. The rest are benchmark names and Spearman ρ correlation scores. The last column shows aggregated Spearman 𝑝 𝑎𝑣𝑔 correlation score of all the benchmarks.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Aggregated Spearman ρ correlation scores over different model types 𝑃 𝑡 .</figDesc><graphic coords="7,181.02,463.74,232.95,179.40" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: The correlation between vector size (d) hyperparameter and benchmark aggregated scores grouped by embedding model type.</figDesc><graphic coords="8,168.00,72.00,258.85,181.45" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: The correlation between vector size (dim) hyperparameter and benchmark aggregated scores. We can use Pearson correlation score 𝑟 (see Eq. 5) to see if there is correlation between values. 𝑟 = 𝑛(∑ 𝑥 𝑦) − (∑ 𝑥)(∑ 𝑦) √ [𝑛 ∑ 𝑥 2 − (∑ 𝑥) 2 ] [𝑛 ∑ 𝑦 2 − (∑ 𝑦) 2 ]</figDesc><graphic coords="8,184.00,343.55,226.55,179.15" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc></figDesc><table><row><cell cols="2">Experimental corpus: initial version</cell><cell></cell><cell></cell></row><row><cell>Corpora</cell><cell>Documents</cell><cell>Token count</cell><cell>Unique tokens</cell></row><row><cell>Wikipedia</cell><cell>286 089</cell><cell>22 942 951</cell><cell>971 506</cell></row><row><cell>CCLL</cell><cell>8 128</cell><cell>136 279 087</cell><cell>2 329 976</cell></row><row><cell>News articles</cell><cell>118 930</cell><cell>60 456 637</cell><cell>718 051</cell></row><row><cell>Total</cell><cell>413 148</cell><cell>219 678 675</cell><cell>2 894 874</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc></figDesc><table><row><cell>Experimental corpus: final version</cell><cell></cell></row><row><cell>Parameter</cell><cell>Value</cell></row><row><cell>Document count</cell><cell>303 443</cell></row><row><cell>Total tokens</cell><cell>160 174 732</cell></row><row><cell>Unique tokens</cell><cell>1 396 607</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 3</head><label>3</label><figDesc>Excerpt of the adapted of SimLex999 benchmark for the Lithuanian language</figDesc><table><row><cell>word1</cell><cell>word2</cell><cell>value</cell><cell>word1</cell><cell>word2</cell><cell>value</cell></row><row><cell>steak</cell><cell>meat</cell><cell>7.47</cell><cell>kepsnys</cell><cell>mėsa</cell><cell>8.2</cell></row><row><cell>nail</cell><cell>thumb</cell><cell>3.55</cell><cell>nagas</cell><cell>nykštys</cell><cell>4.5</cell></row><row><cell>band</cell><cell>orchestra</cell><cell>7.08</cell><cell>grupė</cell><cell>orkestras</cell><cell>7.6</cell></row><row><cell>book</cell><cell>bible</cell><cell>5.00</cell><cell>knyga</cell><cell>biblija</cell><cell>5.6</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 4</head><label>4</label><figDesc>Hyperparameters used for embedding model training</figDesc><table><row><cell></cell><cell>Word2vec</cell><cell>FastText</cell><cell>GloVe</cell></row><row><cell>Architecture Model training</cell><cell>CBOW, Skip-gram Negative Sampling, hierarchical SoftMax</cell><cell>CBOW, Skip-gram Negative Sampling</cell><cell>Global word-word Co-occurrence matrix</cell></row><row><cell>Dimensions</cell><cell>100, 300, 500, 1000</cell><cell>100, 300, 500</cell><cell>100, 300</cell></row><row><cell>Window size</cell><cell>5</cell><cell>5, 10</cell><cell>5, 10</cell></row><row><cell>Minimum count</cell><cell>1, 2, 5</cell><cell>2, 5</cell><cell>2 ,5</cell></row><row><cell cols="4">A total of 50 (19 Word2vec, 20 fastText, 11 GloVe) word embeddings were created by applying</cell></row><row><cell cols="2">different hyperparameter sets.</cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 5</head><label>5</label><figDesc>The best 4 word embeddings ranked by 𝒑 𝒂𝒗𝒈 (Spearman aggregated score)</figDesc><table><row><cell>Model</cell><cell>MEN</cell><cell cols="4">WS353 WS353R WS353S SimLex999</cell><cell cols="3">RG65 MTurk 𝑝 𝑎𝑣𝑔</cell></row><row><cell>FastText SKIP 300d 5w 5m 5i</cell><cell>0.718</cell><cell>0.693</cell><cell>0.539</cell><cell>0.779</cell><cell>0.412</cell><cell>0.733</cell><cell cols="2">0.684 0.651</cell></row><row><cell>FastText SKIP 300d 5w 2m 5i</cell><cell>0.717</cell><cell>0.679</cell><cell>0.513</cell><cell>0.771</cell><cell>0.41</cell><cell>0.737</cell><cell cols="2">0.682 0.644</cell></row><row><cell>FastText SKIP 100d 5w 5m 5i</cell><cell>0.712</cell><cell>0.681</cell><cell>0.544</cell><cell>0.785</cell><cell>0.388</cell><cell>0.721</cell><cell cols="2">0.678 0.644</cell></row><row><cell>Word2vec NSKIP 300d 5w 1m 5i</cell><cell>0.711</cell><cell>0.679</cell><cell>0.507</cell><cell>0.749</cell><cell>0.422</cell><cell>0.766</cell><cell>0.66</cell><cell>0.642</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 6</head><label>6</label><figDesc>The worst 4 word embeddings ranked by 𝒑 𝒂𝒗𝒈 (Spearman aggregated score)</figDesc><table><row><cell>Model</cell><cell cols="5">MEN WS353 WS353R WS353S SimLex999</cell><cell cols="2">RG65 MTurk 𝑝 𝑎𝑣𝑔</cell></row><row><cell>GloVe 300d 10w 1m 5i</cell><cell>0.657</cell><cell>0.584</cell><cell>0.426</cell><cell>0.689</cell><cell>0.378</cell><cell>0.706</cell><cell>0.614 0.579</cell></row><row><cell>GloVe 100d 10w 2m 5i</cell><cell>0.65</cell><cell>0.584</cell><cell>0.414</cell><cell>0.683</cell><cell>0.363</cell><cell>0.724</cell><cell>0.611 0.575</cell></row><row><cell>FastText CBOW 100d 5w 1m 5i</cell><cell>0.65</cell><cell>0.564</cell><cell>0.389</cell><cell>0.679</cell><cell>0.419</cell><cell>0.761</cell><cell>0.559 0.574</cell></row><row><cell>GloVe 100d 10w 1m 5i</cell><cell>0.647</cell><cell>0.588</cell><cell>0.432</cell><cell>0.673</cell><cell>0.356</cell><cell>0.709</cell><cell>0.609 0.573</cell></row></table><note>Comparison between different types of embeddings (Word2vec, fastText, GloVe) was done by averaging 𝑝 𝑎𝑣𝑔 by embedding type (see Eq. (4).</note></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://radimrehurek.com/gensim/models/word2vec.html</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://github.com/facebookresearch/fastText/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">https://github.com/stanfordnlp/GloVe</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<title level="m" type="main">Efficient estimation of word representations in vector space</title>
		<author>
			<persName><forename type="first">Tomas</forename><surname>Mikolov</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1301.3781</idno>
		<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Glove: global vectors for word representation</title>
		<author>
			<persName><forename type="first">J</forename><surname>Pennington</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Socher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Manning</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">EMNLP</title>
		<imprint>
			<biblScope unit="volume">14</biblScope>
			<biblScope unit="page" from="1532" to="1543" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Simlex-999: Evaluating semantic models with (genuine) similarity estimation</title>
		<author>
			<persName><forename type="first">Felix</forename><surname>Hill</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Roi</forename><surname>Reichart</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Anna</forename><surname>Korhonen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computational Linguistics</title>
		<imprint>
			<biblScope unit="volume">41</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="665" to="695" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Multimodal distributional semantics</title>
		<author>
			<persName><forename type="first">Elia</forename><surname>Bruni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Nam-Khanh</forename><surname>Tran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Marco</forename><surname>Baroni</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of artificial intelligence research</title>
		<imprint>
			<biblScope unit="volume">49</biblScope>
			<biblScope unit="page" from="1" to="47" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Word representations: a simple and general method for semi-supervised learning</title>
		<author>
			<persName><forename type="first">Joseph</forename><surname>Turian</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lev</forename><surname>Ratinov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yoshua</forename><surname>Bengio</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 48th annual meeting of the association for computational linguistics</title>
				<editor>
			<persName><forename type="first">T</forename><surname>Mikolov</surname></persName>
		</editor>
		<editor>
			<persName><surname>Chen</surname></persName>
		</editor>
		<meeting>the 48th annual meeting of the association for computational linguistics</meeting>
		<imprint>
			<date type="published" when="2010">2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Efficient estimation of word representations in vector space</title>
		<author>
			<persName><forename type="first">K</forename><surname>Corrado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Dean</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of International Conference on Learning Representations (ICLR)</title>
				<meeting>International Conference on Learning Representations (ICLR)</meeting>
		<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Analogy-based detection of morphological and semantic relations with word embeddings: what works and what doesn&apos;t</title>
		<author>
			<persName><forename type="first">A</forename><surname>Gladkova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Drozd</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Matsuoka</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the NAACL-HLT SRW</title>
				<meeting>the NAACL-HLT SRW<address><addrLine>San Diego, California</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2016-06-12">2016. June 12-17, 2016</date>
			<biblScope unit="page" from="47" to="54" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Don&apos;t count, predict! a systematic comparison of context-counting vs. context-predicting semantic vectors</title>
		<author>
			<persName><forename type="first">Marco</forename><surname>Baroni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Georgiana</forename><surname>Dinu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Germán</forename><surname>Kruszewski</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 52nd Annual Meeting of the Association for Computational Linguistics</title>
				<meeting>the 52nd Annual Meeting of the Association for Computational Linguistics<address><addrLine>V</addrLine></address></meeting>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">The Indo-Europeans: Archeological Problems</title>
		<author>
			<persName><forename type="first">M</forename><surname>Gimbutas</surname></persName>
		</author>
		<idno type="DOI">10.1525/aa.1963.65.4.02a00030</idno>
		<ptr target="https://doi.org/10.1525/aa.1963.65.4.02a00030" />
	</analytic>
	<monogr>
		<title level="j">American Anthropologist</title>
		<imprint>
			<biblScope unit="volume">65</biblScope>
			<biblScope unit="page" from="815" to="836" />
			<date type="published" when="1963">1963</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Dictionary of the Lithuanian Language</title>
		<imprint>
			<date type="published" when="2002-04-19">2002. April 19, 2018</date>
		</imprint>
	</monogr>
	<note>Archived from the original on 2017-08-11</note>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<ptr target="https://www.pingdom.com/blog/the-biggest-and-busiest-languages-on-wikipedia/" />
		<title level="m">The biggest and busiest languages on Wikipedia</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Problems with evaluation of word embeddings using word similarity tasks</title>
		<author>
			<persName><forename type="first">Manaal</forename><surname>Faruqui</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1605.02276</idno>
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">Separated by an un-common language: Towards judgment language informed vector space modeling</title>
		<author>
			<persName><forename type="first">Ira</forename><surname>Leviant</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Roi</forename><surname>Reichart</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1508.00106</idno>
		<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Is Similarity Visually Grounded? Computational Model of Similarity for the Estonian language</title>
		<author>
			<persName><forename type="first">Claudia</forename><surname>Kittask</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Eduard</forename><surname>Barbu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the International Conference on Recent Advances in Natural Language Processing</title>
				<meeting>the International Conference on Recent Advances in Natural Language Processing</meeting>
		<imprint>
			<publisher>RANLP</publisher>
			<date type="published" when="2019">2019. 2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Intrinsic evaluation of Lithuanian word embeddings using WordNet</title>
		<author>
			<persName><forename type="first">Jurgita</forename><surname>Kapočiūtė-Dzikienė</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Robertas</forename><surname>Damaševičius</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="s">Computer Science On-line Conference</title>
		<imprint>
			<date type="published" when="2018">2018</date>
			<publisher>Springer</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Sentiment analysis of lithuanian texts using traditional and deep learning approaches</title>
		<author>
			<persName><surname>Kapočiūtė-Dzikienė</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Robertas</forename><surname>Jurgita</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Marcin</forename><surname>Damaševičius</surname></persName>
		</author>
		<author>
			<persName><surname>Woźniak</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computers</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page">4</biblScope>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">Testing pre-trained Transformer models for Lithuanian news clustering</title>
		<author>
			<persName><forename type="first">Lukas</forename><surname>Stankevičius</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mantas</forename><surname>Lukoševičius</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2004.03461</idno>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">How to generate a good word embedding</title>
		<author>
			<persName><forename type="first">Siwei</forename><surname>Lai</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Intelligent Systems</title>
		<imprint>
			<biblScope unit="volume">31</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="5" to="14" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<idno>1998-2016</idno>
		<ptr target="http://tekstynas.vdu.lt/tekstynas" />
		<title level="m">Kompiuterinės lingvistikos centras. Dabartinės lietuvių kalbos tekstynas</title>
				<imprint/>
	</monogr>
	<note>Vytauto Didžiojo universitetas</note>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<title level="m" type="main">To lemmatize or not to lemmatize: how word normalisation affects ELMo performance in word sense disambiguation</title>
		<author>
			<persName><forename type="first">Andrey</forename><surname>Kutuzov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Elizaveta</forename><surname>Kuzmenko</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1909.03135</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">SEMANTIKA 2</title>
		<ptr target="https://www.vdu.lt/lt/vdu-vykdo-es-finansuojama-projekta-semantika-2/" />
	</analytic>
	<monogr>
		<title level="m">VDU vykdo ES finansuojamą projektą</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">A comparative study of stemming algorithms</title>
		<author>
			<persName><forename type="first">Anjali</forename><surname>Jivani</surname></persName>
		</author>
		<author>
			<persName><surname>Ganesh</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Int. J. Comp. Tech. Appl</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="1930" to="1938" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Multimodal distributional semantics</title>
		<author>
			<persName><forename type="first">Elia</forename><surname>Bruni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Nam-Khanh</forename><surname>Tran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Marco</forename><surname>Baroni</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of artificial intelligence research</title>
		<imprint>
			<biblScope unit="volume">49</biblScope>
			<biblScope unit="page" from="1" to="47" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Placing Search in Context: The Concept Revisited</title>
		<author>
			<persName><forename type="first">Lev</forename><surname>Finkelstein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Evgeniy</forename><surname>Gabrilovich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yossi</forename><surname>Matias</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ehud</forename><surname>Rivlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zach</forename><surname>Solan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gadi</forename><surname>Wolfman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Eytan</forename><surname>Ruppin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Transactions on Information Systems</title>
		<imprint>
			<biblScope unit="volume">20</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="116" to="131" />
			<date type="published" when="2002">2002</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<title level="m" type="main">A study on similarity and relatedness using distributional and wordnet-based approaches</title>
		<author>
			<persName><forename type="first">Eneko</forename><surname>Agirre</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<monogr>
		<title level="m" type="main">A study on similarity and relatedness using distributional and wordnet-based approaches</title>
		<author>
			<persName><forename type="first">Eneko</forename><surname>Agirre</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<monogr>
		<title level="m" type="main">Learning distributed representations of sentences from unlabelled data</title>
		<author>
			<persName><forename type="first">Felix</forename><surname>Hill</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kyunghyun</forename><surname>Cho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Anna</forename><surname>Korhonen</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1602.03483</idno>
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">A word at a time: computing word relatedness using temporal semantic analysis</title>
		<author>
			<persName><forename type="first">Kira</forename><surname>Radinsky</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 20th international conference on World wide web</title>
				<meeting>the 20th international conference on World wide web</meeting>
		<imprint>
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<analytic>
		<title level="a" type="main">Contextual correlates of synonymy</title>
		<author>
			<persName><forename type="first">Herbert</forename><surname>Rubenstein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">John</forename><forename type="middle">B</forename><surname>Goodenough</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Communications of the ACM</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="627" to="633" />
			<date type="published" when="1965">1965</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<monogr>
		<ptr target="https://cloud.google.com/translate" />
		<title level="m">Fast, dynamic translation tailored to your content needs</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
