<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Approaches to assessing the semantic similarity of texts in a multilingual space</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">A</forename><forename type="middle">Kh</forename><surname>Khakimova</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">ANO «Scientific and Research Center for Information in Physics and Technique»</orgName>
								<orgName type="institution">Novgorod</orgName>
								<address>
									<settlement>Nizhny</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">M</forename><forename type="middle">M</forename><surname>Charnine</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">FRC CSC</orgName>
								<orgName type="institution">Russian Academy of Sciences</orgName>
								<address>
									<settlement>Moscow</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Klokov</surname></persName>
							<email>aaklokov@yandex.ru</email>
							<affiliation key="aff1">
								<orgName type="department">FRC CSC</orgName>
								<orgName type="institution">Russian Academy of Sciences</orgName>
								<address>
									<settlement>Moscow</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">E</forename><forename type="middle">G</forename><surname>Sokolov</surname></persName>
							<email>evgeny.sokolov@phystech.edu</email>
							<affiliation key="aff1">
								<orgName type="department">FRC CSC</orgName>
								<orgName type="institution">Russian Academy of Sciences</orgName>
								<address>
									<settlement>Moscow</settlement>
									<country key="RU">Russia</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Approaches to assessing the semantic similarity of texts in a multilingual space</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">0EE15F4532816B63AD1475AAED23C9A3</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T01:46+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>cross-lingual semantic similarity</term>
					<term>semantic textual similarity measure</term>
					<term>semantic implicit links</term>
					<term>collection of documents</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This paper is devoted to the development of a methodology for evaluating the semantic similarity of any texts in different languages is developed. The study is based on the hypothesis that the proximity of vector representations of terms in semantic space can be interpreted as a semantic similarity in the cross-lingual environment. Each text will be associated with a vector in a single multilingual semantic vector space. The measure of the semantic similarity of texts will be determined by the measure of the proximity of the corresponding vectors. We propose a quantitative indicator called Index of Semantic Textual Similarity (ISTS) that measures the degree of semantic similarity of multilingual texts on the basis of identified cross-lingual semantic implicit links. The setting of parameters is based on the correlation with the presence of a formal reference between documents. The measure of semantic similarity expresses the existence of two common terms, phrases or word combinations. Optimal parameters of the algorithm for identifying implicit links are selected on the thematic collection by maximizing the correlation of explicit and implicit connections. The developed algorithm can facilitate the search for close documents in the analysis of multilingual patent documentation.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>As cross-language information retrieval gets more attention, tools to measure cross-language semantic similarity between documents become necessary. An accurate assessment of the actual similarity between documents is fundamental for many automatic text analysis applications, such as thesaurus generation <ref type="bibr" target="#b0">[1]</ref>, machine translation <ref type="bibr" target="#b1">[2]</ref>, information search <ref type="bibr" target="#b2">[3]</ref>, automatic generalization <ref type="bibr" target="#b3">[4]</ref>.</p><p>Text mining and knowledge management technologies play a key role in many areas, including critical infrastructures.</p><p>Information search, document classification, business analytics, forecasting technologies, etc. are currently the most important activities.</p><p>Patent search, including monitoring competitors, checking the novelty of an invention, or searching for technical solutions in other fields of application, requires a lot of effort.</p><p>Comparing documents in different languages is challenging for natural language processing applications, and especially in machine translation applications.</p><p>Cross-language matching of documents is carried out in a patent search to protect an invention in more than one country or region. A separate patent must be filed with several patent offices in different languages. Before applying for a patent, applicants conduct a preliminary search for patents or documents revealing intellectual property similar to the filed invention. In such a process, a set of patents is requested in one language, using the source document in another language as a request.</p><p>To compare the received documents, it is necessary to use cross-language similarity assessment functions. This task can be formulated as discarding text pairs that are not semantically equivalent <ref type="bibr" target="#b4">[5]</ref>. The task is complicated by the fact that in the case of filing an invention in different countries, different standards may be used, which may lead to a discrepancy between versions of the document in different languages. In this case, the task of identifying semantic equivalents is complicated <ref type="bibr" target="#b5">[6]</ref>.</p><p>Natural language processing methods for text analysis and data mining are used in the analysis of many types of technical documentation. Functional analysis methods are based on extracting interactions between the entities described in the document.</p><p>Linguistic analysis tools permit to identify key elements of a document by combining morphological, syntactic, and semantic analysis. Application of methods of linguistic analysis to patent documents allows for accelerated analysis and comparison of patents.</p><p>The purpose of the analysis of technical documentation is to discover possible ambiguities or incompleteness on the one hand, and understanding the requirements in the direction of possible formalization on the other.</p><p>The main problem here is that keyword searches do not take into account synonyms or more abstract terms associated with given query words. This means that if a synonym is used for an important term in a patent application, for example, a wire instead of a cable, a keyword search may not reveal this relationship if an alternative term was not explicitly included in the search query. This is relevant since patent texts often use abstract and general terms to describe the invention in order to maximize protection <ref type="bibr" target="#b6">[7]</ref>.</p><p>If we consider the Internet as a multilingual database, a typical problem when searching for information is the search for relevant documents in the collection of documents by some key terms, or by the example of the corresponding document. Assessing the semantic similarity between words (phrases) is critical to assessing whether a document meets user needs. Many information retrieval systems, such as online library catalog systems, web search engines, deal with multilingual documents and must have tools to measure cross-language semantic similarity.</p><p>In recent decades, many studies have been carried out aimed at improving the effectiveness of measures of semantic similarity of words. However, studies of semantic similarity mainly focus on English. This is partly due to the limited availability of similarity criteria for words in languages other than English. Since the development of multilingual methods is necessary, there is an urgent need to find a reliable basis for assessing multilingual and interlingual semantic similarity.</p><p>Despite the fact that in many areas a multilingual measurement of semantic semantic similarity is required, most algorithms measure semantic similarity between words of the same language. Cross-language similarity was first described in 2009 <ref type="bibr" target="#b7">[8]</ref> for Anglo-Spanish crosslanguage data sets. Over the past few years, multilingual word embeddings, which are lexical elements from several languages in a single semantic space, have attracted considerable attention of researchers <ref type="bibr" target="#b8">[9]</ref><ref type="bibr" target="#b9">[10]</ref><ref type="bibr" target="#b10">[11]</ref>.</p><p>Interlanguage applications are based on data mining methods, such as text clustering, which includes extracting words or phrases from documents as functions, representing documents as feature vectors, and then grouping documents into clusters based on similarity of feature vectors. In a multilingual document collection, recoverable functions will refer to multilingual words. Therefore, it is important to measure the similarity between the words of not only one language, but also of different languages.</p><p>According to the concept of the information data space <ref type="bibr" target="#b11">[12]</ref>, the information space should model a rich set of relationships between data repositories. To model the relationship between data warehouses in data spaces, you need a component that can measure the semantic similarity between interlanguage pairs. Sources in a data space can be relational databases, XML repositories, text databases, web services, etc.</p><p>The problem of plagiarism in a monolingual context is well developed <ref type="bibr" target="#b12">[13]</ref>. Free machine translation tools help spread cross-language plagiarism (plagiarism by translation). In this relatively new field of research, the definition of semantic text similarity in language pairs has been carried out. The authors investigated various existing approaches to detect plagiarism on different language pairs and found that if the method is effective for a particular language pair, it will be equally effective for another language pair with a sufficient number of available lexical resources, i.e. the method can be optimized for a particular case and is effectively applied on another case <ref type="bibr" target="#b13">[14]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Methodology for calculating the assessment of semantic similarity</head><p>The technique includes the following steps: 1) pre-processing of texts by replacing their terms with synset codes; 2) construction of quotation vectors by identifying common rare phrases (long quotes) in various documents using the relevant phrases method; 3) thematic analysis of processed texts and building a set of available topics and corresponding thematic document vectors using the LDA method with the possibility of further clustering documents on topics / ideas into "baskets"/clusters; 4) the construction for each document of an extended vector describing the presence of long citations, the statistics of the synsets included in it and their thematic composition, i.e. the document vector is the concatenation of the citation vector, thematic vector and synset statistics vector; 5) calculation of the similarity index between articles/documents (Semantic Text Similarity Index, ISTS) by the cosine measure of the corresponding article vectors; 6) calculation of the correlation between the formal connectedness of articles and their similarity index, taking into account the minimum and maximum thresholds of the ISTS; 7) the choice of values of various calculation parameters (ISTS thresholds) based on the maximum correlation. The calculation method is selected according to the maximum correlation of ISTS with formal links.</p><p>In the basis of the algorithm for vector transformation of terms used recurrent neural networks (RNN -Recurrent neural network) -Fig. <ref type="figure" target="#fig_0">1</ref>. and IFT, depending on the number of articles with the word in the last 3 years RNN is used for tasks where there is a sequence of words and phrases. Formally, at each step (after each new processed word), RNN considers for each word in the corpus the probability of which word will be next. In this work, LSTM neurons, which are a special case of RNN, were used. Moreover, bi-directional recurrent biLSTM network (Bidirectional recurrent neural networks) was used. biLSTM is a combination of two LSTM networks in which at the same time one network builds a language model from the beginning of the sentence, and the second from the end.</p><p>We used the simplest sequential model, consisting of two layers. For the software implementation of the proposed architecture in Python, the jupyter notebook development environment was used. A linear layer was attached to the biLSTM layer to solve the classification problem (Fig. <ref type="figure">2</ref>).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Fig. 2. Scheme of the LDA model</head><p>At the input of the neural network, vector representations of words (embending) were applied. Word2Vec was used to convert each word from the title of the article to a number vector. In the experiments, 300 dimension vectors were used (Word2Vec from the gensim library allows changing the embedding dimension).</p><p>In our experiments, we consider the DBLP citation network, a collection of articles on artificial intelligence compiled by aminer.org. In this study, we intentionally relied only on the title of the publication and its links. During the experiments, various models of the neural network were tested. Experiments were conducted with a change in the number of neurons in the biLSTM layer (4, 8, 16, 32, 64, 128) and the number of neurons in the linear layer (from 0 to 10). The best model was able to give an accuracy of 0.6131 according to the ROC AUC metric. The time for calculating the forecast and evaluating its accuracy was about 1 hour.</p><p>To combine articles with similar topics into clusters, we used generally accepted approaches to machine word processing (NLP), clustering articles using the Latent Dirichlet Allocation (LDA) method, and visualizing the results obtained with Python libraries. After extracting the data, preprocessing it, extracting tokens, stamping and deleting stop words, we used the Latent Dirichlet Allocation (LDA) algorithm -Fig. <ref type="figure">2</ref>.</p><p>LDA is a hierarchical Bayesian model that consists of two levels: at the first level, a mixture whose components correspond to "themes"; at the second level, a multinomial variable with an a priori Dirichlet distribution that defines the "distribution of topics" in the document.</p><p>The principle of the model: 1) select the document length N 2) vector is selected θ ~ (α) -the vector of the "degree of expression" of each topic in this document; 3) for each of N words w:</p><p>˗ choose a theme zn by distribution Mult(θ); ˗ choose a word wn ~ p(wn|zn, β) with probabilities given in β.</p><p>For simplicity, we fix the number of topics k and assume that β is just a set of parameters βi,j = p(w j =1|z i =1), which need to be evaluated, and we won't worry about the distribution on N. The joint distribution then looks like this:</p><p>Unlike conventional clustering with an a priori Dirichlet distribution, we do not select a cluster here once, and then we look for words from this cluster, but for each word we first select a topic from the distribution θ, and only then we relate this word to this topic.</p><p>At the output after training the LDA model, themed vectors θ are obtained, showing how topics are distributed in each document, and distributions β, which show which words are more likely in certain topics. In our case, we got 8 pronounced clusters corresponding to the following directions: 1) computing systems and algorithms in them; 2) bioinformatics and data processing methods in it; 3) signal processing; 4) optimization methods and algorithms based on them; 5) problems related to theoretical informatics and computational complexity; 6) neural and computing networks; 7) issues regarding natural language processing (NLP) and programming languages; 8) robotics, and self-learning systems (Reinforcement Learning).</p><p>After the previous step, n-dimensional thematic vectors of articles are obtained. To compress the results into a two-dimensional vector space, the t-SNE machine learning algorithm was used. To visualize the clusters, we used an interface written in JavaScrIFT (Fig. <ref type="figure" target="#fig_1">3</ref>).</p><p>The previous approach was based on a comparison of vectors at the megalemma level in a cosine measure, which determined the semantic similarity of the texts. As a development of this approach, based on the assumption that while maintaining the semantic similarity of phrases, ideas in them can be expressed in different words, we use the Impact Factor of the Term (IFT) to assess the similarity of documents.</p><p>To compare articles expressing new ideas, we use the hypothesis that new ideas are often expressed in terms of a high impact factor IFT. IFT is determined by the average number of links to articles with this term, the higher the IFT, the higher the citation trend and the number of formal links. If a couple of articles have a general term with a high IFT, the probability of a formal link between them will be high.</p><p>Using multilingual synsets built for high IFT terms (IFT terms), you can evaluate the similarity of articles in any language. If there is a semantic similarity, estimated by a cosine measure, it can be assumed that articles with this term will be quoted with some probability.</p><p>If previously the similarity of the vectors of megalemma determined the similarity of texts, now we use extended vectors based on common rare phrases, megalemmas and multilingual IFT synsets, as well as the results of thematic analysis. The similarity of extended vectors more accurately reflects the similarity of texts, since it takes into account not only semantic, but also thematic similarity. (purple); 3) signal processing (brown); 4) optimization methods and algorithms based on them (green); 5) problems related to theoretical informatics and computational complexity (orange); 6) neural and computing networks (red); 7) issues regarding natural language processing (NLP) and programming languages (blue); 8) robotics, and self-learning systems (Reinforcement Learning) (dark orange); yellow -a "garbage" cluster with articles in German</p><p>Our study is based on a model for representing ideas in the form of many terms and similar phrases in a multilingual semantic field, on the hypothesis that the proximity of vector representations of terms in a multilingual vector semantic space can be interpreted as semantic similarity in an interlanguage environment. We propose a method of formalizing ideas by using terms with high IFT and megalemma, which allows you to recognize an idea expressed in different words. References, both formal (bibliographic) and contextual (implicit, expressed by matching IFT terms), are an expression of the connection between ideas.</p><p>High IFT terms are significant terms (or ideologically significant). If the texts on the IFT synsets have the same vector, then this means the presence of common ideas in these texts and a significant similarity related to citation. The similarity in vectors of megalemmas also correlates with formal links (as our previous experiments showed), but to a much lesser extent. It is shown that megalemma has a very low impact factor.</p><p>It should be noted that the similarity in vectors of megalemmas is more applicable to texts with common vocabulary, in this case, the degree of coincidence of their thematic composition as a set of popular words is calculated. The approach to calculating the similarity of IFT / megalemma vectors is focused on comparing the similarity of scientific texts with specific terminology, despite the fact that ideas can have different lexical expressions. Therefore, in the second case, it becomes possible to more accurately assess the similarity from the point of view of ideological similarity, since terms with a high IFT are significant terms denoting ideas.</p><p>Three types of semantic similarity can be considered (based on implicit references): 1) similarity of the thematic composition of popular / common words (word frequency from 10 thousand or more); 2) the presence of common significant IFT terms denoting specific ideas (frequency 5-1000); 3) the presence of common rare phrases (long quotation) (frequency 2-100). These types differ in the frequency of matching terms / phrases. The highest frequency is typical for popular terms and megalemmas, the lowest is for common rare phrases. The proposed similarity assessment algorithm takes into account all these types of similarities, giving appropriate weights. Thus, when identifying similarities and implicit references, the entire frequency range of terms and phrases is used.</p><p>So, we build extended vectors from megalemmas and multilingual IFT synsets, and these can be weighted vectors whose elements have weights. The larger the impact factor, the higher the likelihood of a formal link and the higher the weight of the vector element. The cosine measure allows you to work with weighted vectors, in which elements take large real values. Since our task is to search for semantic similarity of articles correlating with the presence of formal links, then increasing the weights of IFT synsets in extended vectors improves the quality of the proposed algorithm.</p><p>Therefore, the algorithm for calculating ISTS is based on assessing the similarity of vectors, expanded by adding multilingual IFT synsets and weights, according to a cosine measure, in order to determine the similarity of texts. This takes into account the presence of formal links between texts containing matching IFT terms. The method may contain options that are determined/selected by the optimization method according to the maximum correlation of ISTS with formal links.</p><p>The first version of the methodology for calculating the multilingual Index of Ideological Influence (III) as the number of similar subsequent / future articles / documents has been developed.</p><p>We consider similar subsequent articles to be articles that will cite this document, i.e. those articles are similar that are linked by formal links. Thus, the III is looking for trending articles containing trending IFT terms. We can calculate the second-level III, since one idea gives rise to another, then you can search for articles similar to the articles found in the first stage (indirect similarity / similarity). The mutual influence of articles is calculated using the PageRank algorithm <ref type="bibr" target="#b14">[15]</ref>, which increases the significance / influence of texts / articles the more they have more (implicit) links with other significant / influential texts.</p><p>IFT terms in scientific articles have an expiration date. The value of IFT is higher in the first years (3-4 years), and then it decreases (Fig. <ref type="figure" target="#fig_2">4</ref>). Over time, some important terms are replaced by others. If in the vector for the term, in addition to the IFT, another year is introduced when the term was of high importance, then you can also obtain some information about the age of the article by the vector, which will allow you to find general ideas of a certain age when comparing the articles. This provides information on the dynamics of the development of ideas. For example, the term NEURAL NETWORKS has a long history, and in different years, various derivatives of this term were significant IFT terms, for example, FUZZY NEURAL or RECURRENT neural networks.</p><p>Over time, some important terms are replaced by others. If in the vector for the term, in addition to the IFT, another year is introduced when the term was of high importance, then you can also obtain some information about the age of the article by the vector, which will allow you to find general ideas of a certain age when comparing the articles. This provides information on the dynamics of the development of ideas. For example, the term NEURAL NETWORKS has a long history, and in different years, various derivatives of this term were significant IFT terms, for example, FUZZY NEURAL or RECURRENT neural networks.</p><p>So, the methodology for calculating the III contains the following steps: 1) search in the article for significant IFT terms; 2) compiling multilingual IFT synsets for these IFT terms; 3) on the basis of IFT-synsets, the definition of the forecast (regression analysis according to previous values of IFT and trend parameters); 4) refinement of the forecast using the PageRank algorithm <ref type="bibr" target="#b11">[12]</ref>, which increases the significance/influence of texts / ideas, the more they have (implicit) connections with other significant / influential texts.</p><p>In this case, implicit links between texts/articles are determined using the methodology for calculating the index of semantic text similarity (ISTS).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Results</head><p>As a result, we see the following pattern: the higher the forecast of the IFT, the higher the III of the document. The predictive value of the IFT is the same as the text, term, or idea. If there are several IFT terms in the text, then you can make a prediction according to the most significant/high IFT, or according to statistics that take into account the synergy of IFT terms when found together. An updated forecast of III/IFT is carried out using regression analysis using a number of indicators for the current year (IFT, IFTm, external links) and similar indicators of previous years.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Conclusion</head><p>The Multilingual Index of Ideological Influence (III) corresponds to the number of subsequent/future articles/documents citing the source document that are similar to the source document. We plan to consider a number of index modifications taking into account the cascade of citation (first and other levels) and the temporal dynamics of the development of ideas. It is planned to develop an algorithm for the updated forecast of III/IFT using a number of indicators of the current year (IFT, IFTm, external links) and similar indicators of previous years.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 .</head><label>1</label><figDesc>Fig. 1. Graphs of the number of points for calculating the correlations of the current and future years for the indicators IFTm (upper) and IFT, depending on the number of articles with the word in the last 3 years</figDesc><graphic coords="2,149.18,524.74,296.34,118.65" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Fig. 3 .</head><label>3</label><figDesc>Fig.3. Cluster states in 1993. 1) computing systems and algorithms in them (pink); 2) bioinformatics and data processing methods in it (purple); 3) signal processing (brown); 4) optimization methods and algorithms based on them (green); 5) problems related to theoretical informatics and computational complexity (orange); 6) neural and computing networks (red); 7) issues regarding natural language processing (NLP) and programming languages (blue); 8) robotics, and self-learning systems (Reinforcement Learning) (dark orange); yellow -a "garbage" cluster with articles in German</figDesc><graphic coords="4,152.88,53.85,289.60,324.59" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Fig. 4 .</head><label>4</label><figDesc>Fig. 4. Graphs of the average values of the IFT term, depending on the number of articles with these terms and the speed of the trend. 1 -0 years, 2 -1 year, 3 -2 years, 4 -3 years, 5 -4 years, 6 -5 years</figDesc><graphic coords="5,86.78,236.68,421.80,278.37" type="bitmap" /></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgment</head><p>The reported study was funded by RFBR according to the research projects № 18-07-00909, 19-07-00857 and 20-04-60185.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Roget&apos;s Thesaurus and Semantic Similarity</title>
		<author>
			<persName><forename type="first">M</forename><surname>Jarmasz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Szpakowicz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Recent Adv. Nat. Lang. Process. III Sel. Pap. from RANLP</title>
		<imprint>
			<biblScope unit="volume">111</biblScope>
			<date type="published" when="2003">2003. 2003. 2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Unsupervised Near-Synonym Choice using the Google Web 1T</title>
		<author>
			<persName><forename type="first">A</forename><surname>Islam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Inkpen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">ACM Trans. Knowl. Discov. Data</title>
		<imprint>
			<biblScope unit="volume">V</biblScope>
			<biblScope unit="page" from="1" to="19" />
			<date type="published" when="2012-06">2012. June</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Semantic matching in search</title>
		<author>
			<persName><forename type="first">H</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Xu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Foundations and Trends in Information Retrieval</title>
		<imprint>
			<biblScope unit="volume">7</biblScope>
			<biblScope unit="issue">5</biblScope>
			<biblScope unit="page" from="343" to="469" />
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">A new sentence similarity measure and sentence based extractive technique for automatic text summarization</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">M</forename><surname>Aliguliyev</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Expert Systems with Applications</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<biblScope unit="page" from="7764" to="7772" />
			<date type="published" when="2009">2009</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Quantifying Cross-lingual Semantic Similarity for Natural Language Processing Applications</title>
		<author>
			<persName><forename type="first">K</forename><surname>Wäschle</surname></persName>
		</author>
		<idno>. -139</idno>
		<imprint>
			<date type="published" when="2015">2015</date>
			<pubPlace>Heidelberg</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Structural and topical dimensions in multi-task patent translation</title>
		<author>
			<persName><forename type="first">K</forename><surname>Wäschle</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Riezler</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 13th Conference of the European Chapter of the Association for Computational Linguistics (EACL)</title>
				<meeting>the 13th Conference of the European Chapter of the Association for Computational Linguistics (EACL)<address><addrLine>Avignon, France</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2012-04-23">2012. April 23 -27, 2012</date>
			<biblScope unit="page" from="818" to="828" />
		</imprint>
	</monogr>
	<note>Proceedings of the 13th Conference of the European Chapter of the Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">The Portability of Three Types of Text Mining Techniques into the Patent Text Genre</title>
		<author>
			<persName><forename type="first">L</forename><surname>Andersson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hanbury</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rauber</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2017">2017</date>
			<publisher>Springer Berlin</publisher>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="241" to="280" />
			<pubPlace>Heidelberg, Berlin, Heidelberg</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">A study on similarity and relatedness using distributional and WordNet-based approaches</title>
		<author>
			<persName><forename type="first">A</forename><surname>Eneko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Enrique</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Keith</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Jana</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Marius</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Aitor</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of Human Language Technologies: The 2009 Annual Conference of the North American Chapter of the Association for Computational Linguistics</title>
				<meeting>Human Language Technologies: The 2009 Annual Conference of the North American Chapter of the Association for Computational Linguistics<address><addrLine>Boulder, Colorado</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2009">2009</date>
			<biblScope unit="page" from="19" to="27" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Bilingual word embeddings for phrase-based machine translation</title>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">Y</forename><surname>Zou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Socher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">M</forename><surname>Cer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of EMNLP</title>
				<meeting>EMNLP</meeting>
		<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="1393" to="1398" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Wiktionary-based word embeddings</title>
		<author>
			<persName><forename type="first">G</forename><surname>De Melo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of MT Summit XV</title>
				<meeting>MT Summit XV</meeting>
		<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="346" to="359" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<title level="m" type="main">Massively multilingual word embeddings</title>
		<author>
			<persName><forename type="first">W</forename><surname>Ammar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Mulcaire</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Tsvetkov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Lample</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Dyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">A</forename><surname>Smith</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1602.01925</idno>
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">From databases to data spaces: A new abstraction for information management</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">F</forename><surname>Michael</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><forename type="middle">H</forename><surname>Alon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>David</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">SIGMOD Record</title>
		<imprint>
			<biblScope unit="volume">34</biblScope>
			<biblScope unit="issue">4</biblScope>
			<biblScope unit="page" from="27" to="33" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Overview of the 6 th International Competition on Plagiarism Detection</title>
		<author>
			<persName><forename type="first">M</forename><surname>Potthast</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Hagen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Beyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Busse</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Tippmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Stein</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">PAN at CLEF 2014</title>
				<meeting><address><addrLine>Sheffield, UK</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="845" to="876" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Using Word Embedding for Cross-Language Plagiarism Detection</title>
		<author>
			<persName><forename type="first">J</forename><surname>Ferrero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Besacier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Schwab</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Agnes</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 15 th Conference of the European Chapter of the Association for Computational Linguistics</title>
				<meeting>the 15 th Conference of the European Chapter of the Association for Computational Linguistics<address><addrLine>Valencia, Spain</addrLine></address></meeting>
		<imprint>
			<publisher>EACL</publisher>
			<date type="published" when="2017">2017. 2017</date>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="415" to="421" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">The PageRank Citation Ranking: Bringing Order to the Web</title>
		<author>
			<persName><forename type="first">L</forename><surname>Page</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Brin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Motwani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Winograd</surname></persName>
		</author>
		<ptr target="evgeny.sokolov@phystech.edu" />
	</analytic>
	<monogr>
		<title level="m">ANO «Scientific and Research Center for Information in Physics and Technique</title>
				<editor>
			<persName><forename type="first">Sokolov</forename><surname>Evgenii</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename></persName>
		</editor>
		<meeting><address><addrLine>Naberezhnye Chelny, Russia; Nizhny Novgorod, Russia; Moscow, Russia; Moscow, Russia; Moscow, Russia</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1998">1998. 1998</date>
		</imprint>
		<respStmt>
			<orgName>Stanford University, Stanford ; FRC CSC of the Russian Academy of Sciences ; FRC CSC of the Russian Academy of Sciences ; FRC CSC of the Russian Academy of Sciences</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical Report</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
