<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">The Implementation of the Mention-Ranking Approach to Coreference Resolution in Russian</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Anna</forename><surname>Kupriianova</surname></persName>
							<email>annkupriyanova26@gmail.com</email>
							<affiliation key="aff0">
								<orgName type="institution">ITMO University Saint-Petersburg</orgName>
								<address>
									<country key="RU">Russian Federation</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Ivan</forename><surname>Shilin</surname></persName>
							<email>shilinivan@corp.ifmo.ru</email>
							<affiliation key="aff0">
								<orgName type="institution">ITMO University Saint-Petersburg</orgName>
								<address>
									<country key="RU">Russian Federation</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Gerhard</forename><surname>Wohlgenannt</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">ITMO University Saint-Petersburg</orgName>
								<address>
									<country key="RU">Russian Federation</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Liubov</forename><surname>Kovriguina</surname></persName>
							<email>lyukovriguina@corp.ifmo.ru</email>
							<affiliation key="aff0">
								<orgName type="institution">ITMO University Saint-Petersburg</orgName>
								<address>
									<country key="RU">Russian Federation</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">The Implementation of the Mention-Ranking Approach to Coreference Resolution in Russian</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">93CE3294BE18AA961E60D473E75F022C</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-23T22:47+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>coreference resolution in Russian</term>
					<term>mention-pair model</term>
					<term>neural network based coreference system</term>
					<term>RuCor</term>
					<term>FastText model</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Coreference resolution is a fundamental ingredient for many downstream tasks in natural language-based applications. For Russian language, the work in coreference resolution is very limited. In this publication, we present a system inspired by the mention-ranking approach, which improves the state-of-the-art F1 score from 0.63 to 0.71, measured with the B 3 metric. We evaluate various sets of feature combinations, and also discuss the limitations of the presented work.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Coreference resolution is an important problem in many natural language processing tasks. It can support, i.e., automatic text summarization, knowledge extraction, objects' identification in dialogue and translation systems <ref type="bibr" target="#b6">[Toldova, Ionov 2017]</ref>. The term coreference denotes the relation between the parts of a text (mentions), that refer to the same real world entities, for example the mention of a person name in a text, and of a pronoun that refers to this same person. Thus, the task of coreference resolution is to find and group all the mentions in the text according to their referents. Mentions are typically represented by noun phrases (NPs), named entities and pronouns, except the cases of abstract anaphora, where the anaphoric pronoun refers to the whole preceding sentence and not to noun phrases (NPs) or pronouns <ref type="bibr" target="#b8">[Nedoluzhko, Lapshinova-Koltunski 2016]</ref>. In the pair of mentions the first one (full mention) is called the antecedent while the second one is an anafor. In the broader field of coreference resolution, there are two main tasks: mention extraction and coreference resolution in narrower sense (mention clustering) <ref type="bibr" target="#b7">[Sysoev et al. 2017]</ref>. Mention extraction finds textual expressions which are possible elements of coreference chains in unstructured data, whereas coreference resolution groups mentions into clusters, which refer to a single real-world entity. The system presented here focuses on the second task, i.e. coreference resolution in the narrower sense.</p><p>For Russian language, the work on coreference resolution is limited. Results reported on the RuCor<ref type="foot" target="#foot_0">1</ref>  <ref type="bibr" target="#b5">[Toldova et al. 2014</ref>] coreference corpus are 60.48 of F1 score for the B 3 metric by <ref type="bibr" target="#b6">Toldova and Ionov [Toldova, Ionov 2017]</ref>, and 63.12 F1 by <ref type="bibr" target="#b7">Sysoev et al. [Sysoev et al. 2017</ref>]. The mentioned work applies rule-based and "classical" machinelearning methods like decision trees or logistic regression. In this work, we present an approach using neural networks based on an adapted version of the mention-ranking model by <ref type="bibr" target="#b4">Clark and Manning [Clark, Manning 2016]</ref>. With this architecture, we manage to outperform previous work with an achieved F1-score of 0.7131. The B 3 metric <ref type="bibr" target="#b9">[Bagga, Baldwin 1998</ref><ref type="bibr" target="#b10">, Amigo et al. 2009</ref>] is a clustering metric, which evaluates a gold-standard clustering of mentions against a system-produced clustering.</p><p>The paper is structured as follows: After an overview of related work in Section 2, Section 3 introduces the system architecture, and Section 4 discusses the features used, and how features are combined into three different sets. The following section (Section 5) provides the evaluation details for those feature sets and compares the results to the state-of-the-art. Furthermore, difficult cases are discussed. The paper concludes with Section 6.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Related Work</head><p>Existing approaches to coreference resolution can be divided into heuristic <ref type="bibr" target="#b0">[Hobbs 1978</ref><ref type="bibr" target="#b1">, Boyarski et al. 2013</ref> ] and based on machine learning (ML) algorithms <ref type="bibr" target="#b2">[Rahman, Ng 2009</ref><ref type="bibr" target="#b3">, Ng 2008</ref><ref type="bibr" target="#b4">, Clark, Manning 2016]</ref>. Heuristic methods are built upon a handmade set of rules, which is time-and labour-consuming to construct. On the contrary, ML-based approaches are faster and easier to develop, but they depend on the availability of a coreference dataset of sufficient size and quality to apply supervised learning methods.</p><p>Recent advances in coreference resolution for English language include the work of Clark and Manning <ref type="bibr" target="#b4">[Clark, Manning 2016</ref>] on a cluster-ranking algorithm that handles entity-level information and eliminates the disadvantages of the mention-pair models. The main benefit of this neural network-based method is that it can distinguish beneficial cluster merges from harmful ones. Central parts of the system architecture used for this publication are inspired by this work.</p><p>In general, the state-of-the-art results for Russian language are lower than for English, work on Russian language is rather limited so far <ref type="bibr" target="#b7">[Sysoev et al. 2017]</ref>. For Russian language, coreference resolution became a more active research topic with the release of a tagged coreference corpus (RuCor) in 2014 <ref type="bibr" target="#b5">[Toldova et al. 2014</ref>]. <ref type="bibr" target="#b6">Toldova and Ionov [Toldova, Ionov 2017]</ref> compare rule-based and ML-based methods for coreference resolution using this RuCor dataset, with slightly better performance for the ML-based methods. They reach 31.56 for predicted mentions and 60.48 for gold mentions with the B 3 metric for the RuCor corpus. Predicted mentions refers to coreference resolution for mentions which were automatically extracted with a mention extraction module. Sysoev et al. <ref type="bibr" target="#b7">[Sysoev et al. 2017]</ref> tackle both mention extraction and coreference resolution. For mention extraction, they use a number of linguistic, structural, etc., features and apply classifiers such as logistic regression, Jaccard Item Set mining and random forest, and reach an F1 of 63.12 for the gold mentions from the RuCor dataset. In comparison, our approach provides an F1-score about 8 points above previous work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">System Architecture</head><p>We present a coreference resolution system based on the mention-ranking model by Clark and Manning <ref type="bibr" target="#b4">[Clark, Manning 2016]</ref>. The core of the system is a feedforward neural network. Its topology is shown in Figure <ref type="figure" target="#fig_1">1</ref>. In a nutshell, the network can be divided into two parts: a mention-pair encoder and a mention-ranking model. The implementation of the system is available on github<ref type="foot" target="#foot_1">2</ref> . The source code was written in Python, using Keras and Tensorflow for the neural network models.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Input Layer h 0</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Hidden Layer h 1</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Hidden Layer h 3</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Hidden Layer h 2</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Output Layer s m</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>ReLU (W</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1">Mention-Pair Encoder</head><p>The purpose of the mention-pair encoder is to transform a pair of a mention m and its potential antecedent a into their distributed representations. The mention-pair encoder is implemented as a feedforward neural network with three fully-connected hidden layers of rectified linear units (ReLU):</p><formula xml:id="formula_0">h i (a, m) = max(0, W i h i−1 (a, m) + b i ) (1)</formula><p>where h i (a, m) is an output of the i-th hidden layer for a pair of mention m and its potential antecedent a. W i is a weight matrix and b i is the bias for the i-th hidden layer.</p><p>The input layer of the mention-pair encoder takes a vector of features of a mention and its potential antecedent as well as additional pair features (all the features are described in Section 4). The output of the last hidden layer is the distributed representation of the pair which is used as an input to the mention-ranking model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2">Mention-Ranking Model</head><p>The purpose of the mention-ranking model is to estimate the score of coreference compatibility for the pair of a mention m and its potential antecedent a. To compute this score one applies one fully-connected layer to the distributed representation of the pair</p><formula xml:id="formula_1">r m (a, m), s m (a, m) = W m r m (a, m) + b m (2)</formula><p>where s m (a, m) denotes a score of coreference compatibility of a pair of mention m and its potential antecedent a, r m (a, m) is the distributed representation of this pair.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3">Training objective</head><p>For pretraining, which determines the initial configuration of the model parameters, we used the following objective function,</p><formula xml:id="formula_2">− N i=1 t∈T (m i ) log p(t, m i ) + f ∈F (m i ) log(1 − p(t, m i ))<label>(3)</label></formula><p>where T (m i ) and F (m i ) are sets of true and false antecedents of a mention m i respectively, and p(a, m i ) = sigmoid(s(a, m i )).</p><p>The main training objective is a slack-rescaled max-margin which penalizes different types of errors:</p><formula xml:id="formula_3">N i=1 max a∈A(m i ) ∆(a, m i )(1 + s m (a, m i ) − s m ( ti , m i ))<label>(4)</label></formula><p>where A(m i ) is a set of candidate antecedents of a mention m i , ti is a highest scoring true antecedent of mention m i : ti = arg max</p><formula xml:id="formula_4">t∈T (m i ) s m (t, m i )<label>(5)</label></formula><p>and ∆(a, m i ) is a cost function for different types of mistakes:</p><formula xml:id="formula_5">∆(a, m i ) = α F N if a = N A ∧ T (m i ) = N A α F A if a = N A ∧ T (m i ) = N A α W L if a = N A ∧ a / ∈ T (m i ) 0 if a ∈ T (m i )<label>(6)</label></formula><p>where F N stands for False New mistake, F A for False Anaphor, and W L for Wrong Link.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Feature Sets and Models</head><p>For creating the coreference resolution models, we designed three feature sets. We will compare the results for the individual models in the evaluation section. Table <ref type="table" target="#tab_0">1</ref> shows how the features are partitioned into our feature sets.</p><p>The list of features is inspired by previous work on English and Russian coreference resolution. The feature set I is a reduced version of the features used by Clark and Manning <ref type="bibr" target="#b4">[Clark, Manning 2016]</ref>. In feature set II we removed some of the word embedding features, and added features of explicit indication of morphological characteristics and POS-tag agreement, which is highly relevant for Russian as a morphologically rich language<ref type="foot" target="#foot_2">3</ref> . And in feature set III the cosine similarity between the vectors of mentions heads completely replaces the word embeddings.</p><p>For each mention, we build a feature vector. A mention consisting of one word, is represented by a single vector (word embedding). If the mention consists of a group of words, is represented by the average of the embedding vectors of each word in the group. We use word embeddings pre-trained with FastText on the Wikipedia corpus<ref type="foot" target="#foot_3">4</ref> .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Experiments and Results</head><p>This section describes the experimental setup, especially the dataset, and provides the results of the evaluations for the three models introduced in Section 4 in comparison with existing work. Finally, we discuss some of the difficulties and limitations observed with the current architecture.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1">Experimental Setup</head><p>With regards to the dataset, we used the RuCor dataset to train and test the coreference resolution system. RuCor is the first open corpus with coreference annotations for the Russian language. It comprises short texts and text fragments of different genres: news, fiction, scientific papers, etc. All the texts are tokenized, split into sentences and parsed syntactically and morphologically. In the corpus, mentions are limited to NPs that refer to real-world entities. Thus, abstract and generic NPs as well as bridging relations and coreference relations with a split antecedent are not annotated. The RuCor dataset contains 181 texts with 156637 tokens and 3638 coreference chains. Based on these numbers, in model training we used a 60-20-20 percent split to create the training, validation, and test datasets.</p><p>In terms of the settings of the neural network models, we minimized the training objectives with the Adam and RMSProp optimizers. For regularization, we applied dropout with a rate of 0.3 on the output of each hidden layer.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.2">Experiments</head><p>For the feature sets three models of the neural network were developed. All the models were pretrained. Pretraining is performed for an initial setup of the weights, and uses binary cross-entropy to compute the loss when training only the mention-ranking model. The results of the experiments are shown in Table <ref type="table" target="#tab_1">2</ref>. According to Table <ref type="table" target="#tab_1">2</ref> Model II shows the best results. We suppose that word embeddings and explicit indication of the matches in morphological attributes and POS-tags positively influence the results. In Model III word embeddings are replaced with cosine similarity between the embeddings for the mentions' head words. It appears that is may not be sufficient for the identification of their semantic similarity. Model I has the lowest score which might be explained by the absence of the explicit indication of matches in features, which, as stated above, lowers the model quality. Moreover, the non-proportional lengths of the vectors for different features, with 300-dimensional vectors for the head word of each mention in a pair and all the words of these mentions, in contrast with only 38-dimensional vector for the other features, might influence the outcome.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.3">Comparison</head><p>In Table <ref type="table" target="#tab_2">3</ref> we compare the results of our system with the state-of-the-art open coreference resolution systems for Russian by Tolodova and Ionov <ref type="bibr" target="#b6">[Toldova, Ionov 2017</ref>] and Sysoev et al. <ref type="bibr" target="#b7">[Sysoev et al. 2017]</ref>. We compare the B 3 metric on the gold mentions, i.e. mentions from the RuCor corpus (not for mentions automatically extracted from the text). All our models surpass existing work with regards to F1 score, with model II giving the best results. The comparison baselines can be briefly described as follows (for details see Tolodova and Ionov <ref type="bibr" target="#b6">[Toldova, Ionov 2017]</ref>, and Sysoev et al. <ref type="bibr" target="#b7">[Sysoev et al. 2017]</ref>): The MLUpdated model implements a ML-based decision tree classifier. The NamedEntities model takes into account semantic information in the form of the lists of possible named entities. This allows it to compare the mentions' semantic classes. The Word2vec model uses word embeddings for evaluating the semantic compatibility of the mentions' heads (we used the same feature in our Model I and Model II). <ref type="bibr" target="#b7">Sysoev et al. [Sysoev et al. 2017</ref>] use a common set of features, which is fed into various classifiers such as logistic regression with Jaccard Item set mining, or a random forest.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.4">Error Analysis and Results Discussion</head><p>Here, we outline some of the problems and errors which have been discovered in the analysis of the predictions of the neural network:</p><p>1. Some errors are caused by the wrong annotations in the RuCor coreference corpus, esp. wrong lemmas or morphological attributes in the corpus. For example, for the word "дотком" (dotcom) two different lemmas were found -"доткома" and "доткомом".</p><p>2. Direct speech mistakes: Pronouns "я" (I) and "ты" (you), if used in a dialogue by different speakers, can be coreferential in a certain context. For example, in case of the following dialogue:</p><p>-"Я сегодня выполнил работу за два дня." (Today I have done the work for two days.)</p><p>-"Ты -молодец! " (You did well!)</p><p>The pronouns "я" (I) and "ты" (you) have the same referent -the first speaker. However, the neural network makes this kind of mistakes because there is no information about speakers in the dataset.</p><p>3. Context mistakes: They arise when the coreference relation gets evident only after the analysis of the mentions context. For example, the coreference relation between the mentions "Их Сиятельство" (Their Majesty / Highness) and "женщина в черном капоте" (the woman wearing black dressing gown) is not clear without the analysis of the context parts of the text. Such types of mistakes can be explained with the difficulty of formalizing semantic information. One possible solutions for this problem might be the use of word embeddings for a longer context or even for the whole text.</p><p>4. Split anafor mistakes: For example, the mentions "они" (they) and "Иван Тихонович и Татьяна Финогеновна" (Ivan Tihonovich and Tatyana Finogenovna) are coreferential. But the network makes the mistake because in the dataset the morphological attributes (gender, number, animacy, etc.) are identified only for the head elements of the mentions. And in the above stated case there are two heads in the second mention and they differ in gender with each other and in number with the head of the first mention.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6">Conclusion</head><p>We have presented a coreference resolution system implementing the mention-ranking approach, and experimented with different sets of feature combinations and evaluated their impact on system quality with the B 3 metric. Our best model provides an F1 score of 0.7131 on the RuCor dataset, and exceeds existing work by around 8 points. In future work, there are a number of directions to improve model quality: (i) experimenting with other network architectures, (ii) hyperparameter tuning, and (iii) adding more relevant features into the training process, and finally (iv) applying a cluster-ranking approach to capture additional entity-level information. Furthermore, we plan to extend our system with a mention extraction module.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Neural Network Topology</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>Feature Sets</figDesc><table><row><cell>Feature</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc></figDesc><table><row><cell></cell><cell cols="3">: Experiment Results</cell><cell></cell></row><row><cell></cell><cell>AUC</cell><cell></cell><cell>B 3 metric</cell><cell></cell></row><row><cell></cell><cell></cell><cell cols="3">Precision Recall F1 score</cell></row><row><cell></cell><cell></cell><cell>Training</cell><cell></cell><cell></cell></row><row><cell>Model I</cell><cell>0.6538</cell><cell>0.6397</cell><cell>0.6711</cell><cell>0.6550</cell></row><row><cell cols="5">Model II 0.8280 0.7170 0.7092 0.7131</cell></row><row><cell cols="2">Model III 0.7870</cell><cell>0.6783</cell><cell>0.6902</cell><cell>0.6842</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc></figDesc><table><row><cell cols="2">: Comparison</cell><cell></cell><cell></cell></row><row><cell>Model</cell><cell></cell><cell>B 3 metric</cell><cell></cell></row><row><cell></cell><cell cols="3">Precision Recall F1 score</cell></row><row><cell>Model I</cell><cell>0.6397</cell><cell>0.6711</cell><cell>0.6550</cell></row><row><cell>Model II</cell><cell>0.7170</cell><cell cols="2">0.7092 0.7131</cell></row><row><cell>Model III</cell><cell>0.6783</cell><cell>0.6902</cell><cell>0.6842</cell></row><row><cell>Toldova and Ionov [Toldova, Ionov 2017]:</cell><cell></cell><cell></cell><cell></cell></row><row><cell>MLUpdated</cell><cell>0.7937</cell><cell>0.4860</cell><cell>0.6029</cell></row><row><cell>Toldova and Ionov [Toldova, Ionov 2017]:</cell><cell></cell><cell></cell><cell></cell></row><row><cell>NamedEntities</cell><cell>0.7937</cell><cell>0.4886</cell><cell>0.6048</cell></row><row><cell>Toldova and Ionov [Toldova, Ionov 2017]:</cell><cell></cell><cell></cell><cell></cell></row><row><cell>Word2vec</cell><cell>0.7925</cell><cell>0.4864</cell><cell>0.6028</cell></row><row><cell>Sysoev et al. [Sysoev et al. 2017]:</cell><cell></cell><cell></cell><cell></cell></row><row><cell>log. regr. + Jaccard Item Set mining</cell><cell>0.6014</cell><cell>0.6103</cell><cell>0.6055</cell></row><row><cell>Sysoev et al. [Sysoev et al. 2017]:</cell><cell></cell><cell></cell><cell></cell></row><row><cell>random forest</cell><cell>0.7389</cell><cell>0.5516</cell><cell>0.6312</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">http://rucoref.maimbava.net/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://github.com/annkupriyanova/Coreference-Resolution</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">Morphological annotation, lemmatization and word embedding models were borrowed from<ref type="bibr" target="#b10">[Kovriguina et al. 2017]</ref> </note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">https://github.com/facebookresearch/fastText/blob/master/pretrained-vectors.md</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>L.Kovriguina acknowledges support from the Russian Fund of Basic Research (RFBR), Grant No. 16-36-60055. Furthermore, the work is supported by the Government of the Russian Federation (Grant 074-U01) through the ITMO Fellowship and Professorship Program.</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Resolving pronoun references</title>
		<author>
			<persName><forename type="middle">J</forename><surname>Hobbs ; Hobbs</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Lingua</title>
		<imprint>
			<biblScope unit="volume">44</biblScope>
			<biblScope unit="page" from="311" to="338" />
			<date type="published" when="1978">1978. 1978. 1978</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Viyavlenie anaforicheskih otnosheni pri avtomaticheskom analize teksta [Identification of anaphoric relations in automatic text analysis</title>
		<author>
			<persName><surname>Boyarski</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Выявление анафорических отношений при автоматическом анализе текста. //Научно-технический вестник информационных технологий, механики и оптики</title>
				<editor>
			<persName><forename type="first">Е</forename><forename type="middle">А</forename><surname>Каневский</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">А</forename><surname>Степукова</surname></persName>
		</editor>
		<imprint>
			<date type="published" when="2013">2013. 2013. 2013. 2013</date>
			<biblScope unit="page" from="108" to="112" />
		</imprint>
	</monogr>
	<note>//Nauchno-tehnicheski vestnik informacionnyh tehnologi, mehaniki i optiki</note>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Supervised models for coreference resolution</title>
		<author>
			<persName><surname>Rahman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ng ; Rahman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Ng</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2009 conference on empirical methods in natural language processing</title>
				<meeting>the 2009 conference on empirical methods in natural language processing<address><addrLine>Singapore</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2009">2009. 2009. 2009</date>
			<biblScope unit="page" from="968" to="977" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Unsupervised models for coreference resolution</title>
		<author>
			<persName><forename type="first">;</forename><surname>Ng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Ng</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2008 conference on empirical methods in natural language processing</title>
				<meeting>the 2008 conference on empirical methods in natural language processing<address><addrLine>Honolulu</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2008">2008. 2008. 2008</date>
			<biblScope unit="page" from="640" to="649" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Improving Coreference Resolution by Learning Entity-Level Distributed Representations</title>
		<author>
			<persName><forename type="first">Manning ;</forename><surname>Clark</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Clark</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="s">Association for Computational Linguistics Proceedings</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="643" to="653" />
			<date type="published" when="2016">2016. 2016. 2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Evaluating Anaphora and Coreference Resolution for Russian</title>
		<author>
			<persName><surname>Toldova</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">//Computational Linguistics and Intellectual Technologies</title>
				<imprint>
			<publisher>RGGU</publisher>
			<date type="published" when="2014">2014. 2014. 2014</date>
			<biblScope unit="page" from="681" to="695" />
		</imprint>
	</monogr>
	<note>DIALOG 2014</note>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Coreference Resolution for Russian: The Impact of Semantic Features</title>
		<author>
			<persName><forename type="first">Ionov</forename><forename type="middle">;</forename><surname>Toldova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Toldova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ionov</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Computational Linguistics and Intellectual Technologies. International Conference &quot;Dialog 2017&quot; Proceedings</title>
				<imprint>
			<date type="published" when="2017">2017. 2017. 2017</date>
			<biblScope unit="page" from="339" to="349" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Coreference Resolution in Russian: State-of-the-Art Approaches Application and Evolvement. //Computational Linguistics and Intellectual Technologies</title>
		<author>
			<persName><surname>Sysoev</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference</title>
				<imprint>
			<date type="published" when="2017">2017. 2017</date>
			<biblScope unit="volume">16</biblScope>
			<biblScope unit="page" from="327" to="347" />
		</imprint>
	</monogr>
	<note>Dialog 2017</note>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Abstract Coreference in a Multilingual Perspective: a View on Czech and German</title>
		<author>
			<persName><forename type="first">Lapshinova-Koltunski ;</forename><surname>Nedoluzhko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Nedoluzhko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Lapshinova-Koltunski</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Workshop on Coreference Resolution Beyond OntoNotes (CORBON</title>
				<meeting>the Workshop on Coreference Resolution Beyond OntoNotes (CORBON</meeting>
		<imprint>
			<date type="published" when="2016">2016. 2016. 2016</date>
			<biblScope unit="page" from="47" to="52" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Entity-based Cross-document Coreferencing Using the Vector Space Model</title>
		<author>
			<persName><forename type="first">Baldwin</forename><forename type="middle">;</forename><surname>Bagga</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bagga</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Baldwin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of ACL&apos;98</title>
				<meeting>ACL&apos;98<address><addrLine>Quebec, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1998">1998. 1998. 1998. Montreal</date>
			<biblScope unit="page" from="79" to="85" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">A comparison of Extrinsic Clustering Evaluation Metrics based on Formal Constraints //Information Retrieval</title>
		<author>
			<persName><surname>Amigo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Knowledge Engineering and the Semantic Web</title>
				<imprint>
			<publisher>-Springer</publisher>
			<date type="published" when="2009">2009. 2009. 2009. 2017. 2017. 2017</date>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="page" from="101" to="111" />
		</imprint>
	</monogr>
	<note>Russian Tagging and Dependency Parsing Models for Stanford CoreNLP Natural Language Toolkit</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
