<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Identification of singleton mentions in Russian</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Max Ionov</string-name>
          <email>max.ionov@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Svetlana Toldova</string-name>
          <email>toldova@yandex.ru</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Goethe University Frankfurt / Moscow State University</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>National Research University “Higher School of Economics”</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>Аннотация This paper describes a pilot study of the problem of detecting singleton mentions in Russian texts. A noun phrase is considered a singleton mention if it is the only referent of some entity. We discuss various morphosyntactic and lexical features, some of which were used for analogous tasks for English and propose new features derived from the discourse analysis. Testing the machine learning classifiers trained with the use of proposed features, we conclude that although the quality of classifiers is significantly lower than for English, they still have rather high precision and thus can be helpful in various tasks of mention tracking.</p>
      </abstract>
      <kwd-group>
        <kwd>coreference resolution</kwd>
        <kwd>mention detection</kwd>
        <kwd>discourse processing</kwd>
        <kwd>natural language processing</kwd>
        <kwd>machine learning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        Coreference resolution is an important preprocessing step for many advanced
NLP tasks that deal with discourse processing. Although this task has been
actively researched for several decades (e.g. [
        <xref ref-type="bibr" rid="ref10 ref7">11,8</xref>
        ]), there is still a substantial
amount of work going on in this direction. One of the problems that received
much attention in the last years is the improvement of an important preprocessing
step mention detection.
      </p>
      <p>
        Coreference resolution task can be seen as a task of linking pairs of mentions
that refer to the same discourse entity. However it has been shown ([
        <xref ref-type="bibr" rid="ref12">13</xref>
        ]) that not
every noun phrase in a text likes to be linked, some of them appear just once,
henceforth referred to as singletons. Moreover, the fraction of such mentions
is very high (about 50%) and, as a consequence of it, the chance of linking a
singleton with something by mistake is relatively high. It was demonstrated
(ibid) that filtering out singletons before coreference resolution improves its
overall quality. In addition, removing a large number of unnecessary noun phrases
should improve the computational efficiency of the system. Other high-level
tasks that require tracking of discourse mentions can also benefit from singleton
detection.
      </p>
      <p>In this paper we conduct a pilot study on singleton detection for Russian.
We employ the features that were applied to the similar task for English and
check if they work for Russian. In addition we discuss some new, Russian specific
features. In the experiment we test different subsets of them and measure the
importance of each feature.</p>
      <p>The rest of the paper is structured as follows. In section 2 we give an overview
of the previous work on singleton detection and some theoretical literature on
discourse that deals with certain peculiarities of singleton mentions. In section 3
we describe the corpus that we used for our experiments, the set up for the
experiments and their results. In section 4 we analyze the impact of the features
and discuss expected and unexpected outcomes.
2
2.1</p>
    </sec>
    <sec id="sec-2">
      <title>Background and motivation</title>
      <sec id="sec-2-1">
        <title>Singleton detection in English</title>
        <p>
          Among the vast amount of research devoted to coreference resolution for English,
there has been some work done on various aspects of mention tracking, e.g.
discourse-new detection ([18]) finding NPs that start coreferential chains;
distinguishing anaphoric and non-anaphoric NPs ([
          <xref ref-type="bibr" rid="ref8">9</xref>
          ]) detecting definite NPs
whose interpretation does not depend on previous mentions. Another closely
related type of mention tracking is singleton detection ([
          <xref ref-type="bibr" rid="ref12">13</xref>
          ]) detecting NPs
that correspond to the entities that are mentioned only once.
        </p>
        <p>
          The definiteness category is usually an important factor for those tasks, and
some features discussed in the papers rely crucially on the presence of specific
articles (see [
          <xref ref-type="bibr" rid="ref8">9</xref>
          ], for example). Other researchers employ more complicated features
based on determiners. [18] mentions that unique mentions are used with the
definite article much more often than with indefinite one, like the singleton “the
government” in her corpus was found in 23.9% usages while its “a government”
version only in 4.8%. As the result, for every unique noun phrase, the “definiteness
probability” was assessed and used as a feature in the further classification.
        </p>
        <p>Since Russian is an article-less language, we could not use features based on
the overt morphosyntactic definiteness.</p>
        <p>
          The present work is mainly based on Recasens et al. ([
          <xref ref-type="bibr" rid="ref12">13</xref>
          ]) where the impact
of different morphosyntactic and semantic features on the singleton vs. repetitive
mention detection was analysed. To estimate this impact authors used a binary
logistic regression model. The feature set used for it relied on the theoretical
assumptions made in the theoretical literature on the referential choice modeling.
According to Prince [
          <xref ref-type="bibr" rid="ref11">12</xref>
          ], [1] or Centering Theory ([
          <xref ref-type="bibr" rid="ref4">5</xref>
          ]), the morphosyntactic
properties of a target NP could be relevant. Besides, different types of NPs have
different probability of being coreferent. For instance, anaphoric pronouns are
very likely to be coreferent, the animacy, number and some other properties of
an NP correlate with the likelihood of it being linked to other NPs in the text.
According to [
          <xref ref-type="bibr" rid="ref12">13</xref>
          ] indefinite or negative pronouns have negative association with
chaining.
        </p>
        <p>The other cluster of features discussed there is based on information structure,
most importantly, syntactic position. Subjects as well as verb arguments have
positive association with coreferent use of an NP.
2.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>Features for non-repeating mentions in Russian</title>
        <p>
          As for Russian, there is a lack of research concerning the unique or non-coreferring
expressions properties. However, there are works devoted to special devices for
new salient entities introduction into discourse (e.g. [
          <xref ref-type="bibr" rid="ref1">2</xref>
          ], [
          <xref ref-type="bibr" rid="ref3">4</xref>
          ], [
          <xref ref-type="bibr" rid="ref14">15</xref>
          ], and [
          <xref ref-type="bibr" rid="ref15">16</xref>
          ] for
discourse-new detection).
        </p>
        <p>
          Among the features used there is NP length (the first NP for a new salient
referent tends to be longer) and, in particular, the number of adjectives. For
instance, [
          <xref ref-type="bibr" rid="ref14">15</xref>
          ] suggests the following examples:
(1)
a. ? On podoˇsol k stolu, vzjal s nego ˇcernuju ruˇcku v forme kinˇzala, ?sel
i stal pisat’.
        </p>
        <p>He came up to the table, took a black pen in the shape of a dagger
from it ?and started writing.
b. On podoˇsol k stolu, vzjal ruˇcku, sel i stal pisat’.</p>
        <p>He came up to the table, took a pen and started writing.</p>
        <p>In both examples the underlined entity (the pen) is a singleton. In the example 1a
it has two modifiers and the end of the sentence without mentioning the entity
again sounds unnatural. In the example 1b the NP is a typical (‘expected’)
participant of the ‘writing’ event. Therefore, the sentence seems acceptable.</p>
        <p>
          According to [
          <xref ref-type="bibr" rid="ref15">16</xref>
          ], there are also some specific lexical features, for example,
so-called non-identity words like takoj (such) and some adjectives have an impact
on the status of an NP in the discourse.
        </p>
        <p>Thus, the question is whether those features that showed their usefulness in
the first-mentions detection task, could be helpful for the singleton differentiation
as well.
3
3.1</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Experiments</title>
      <sec id="sec-3-1">
        <title>Data</title>
        <p>
          Our experiments were conducted on RuCor, a Russian coreference corpus released
as a part of the RU-EVAL campaign ([
          <xref ref-type="bibr" rid="ref16">17</xref>
          ])3.
        </p>
        <p>
          The corpus consists of short texts or fragments of texts in a variety of genres:
news, scientific articles, blog posts and fiction. The whole corpus contains about
180 texts and 3 638 coreferential chains with 16 557 noun phrases in total. Each
text in the corpus is tokenized, split into sentences and morphologically tagged
using tools developed by Serge Sharoff ([
          <xref ref-type="bibr" rid="ref13">14</xref>
          ]). For the experiments described in
this paper, texts were additionally syntactically parsed using the same tools.
The morphological tags were checked and fixed manually, since it was previously
shown that errors on this level affects significantly the quality of a related task
([
          <xref ref-type="bibr" rid="ref6">7</xref>
          ]). The corpus was randomly split into a training and a test set (70% and 30%
respectively).
3 The corpus may be downloaded on http://rucoref.maimbava.net.
        </p>
        <p>
          Since the RuCor annotation followed MUC guidelines ([
          <xref ref-type="bibr" rid="ref5">6</xref>
          ]), singletons are
not annotated, so every unannotated noun phrase was considered a singleton.
This means that we do not distinguish mentions that are never coreferent and
potentially coreferent mentions used only once in a text; even though they may,
in principle, have different structural properties.
3.2
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>Features for singleton detection</title>
        <p>
          In this pilot study, we tested 4 groups of features: basic, structural, lexical,
and syntactic features. Most of the features that we used were proposed before
for detecting singleton mentions in English (e.g. [
          <xref ref-type="bibr" rid="ref12 ref8">13,9</xref>
          ]). Some other features,
correlated with entity discourse role, were previously used in the first mention
detection task ([
          <xref ref-type="bibr" rid="ref15">16</xref>
          ]).
        </p>
        <p>
          As we mentioned already, our notion of singletons collapses two types of
mentions: those that can not be anaphoric and those that were mentioned only
once in the text. In order to detect both of them, we compiled features that detect
non-anaphoricity with those that should have correlation with the discourse role:
non-coreferent mentions should be less important for the discourse.
Basic features The most basic feature is the number of occurrences of the NP
in question or its head in the text before. It is obvious that if an NP is repeated,
chances are that the entity is not a singleton. Other basic features check whether
the noun phrase is animate, a proper noun, contains non-cyrillic characters or is
a pronoun. Those features were shown to be useful for English in [
          <xref ref-type="bibr" rid="ref12">13</xref>
          ].
Structural features This group contains two features: NP length (in words)
and the number of adjectives. Both of them should correlate with the entity
importance in the discourse: the more important an entity is, the more words
would be spent on it. Those features showed a great impact on the first mention
detection task ([
          <xref ref-type="bibr" rid="ref15">16</xref>
          ]), having a strong correlation with the discourse role of a
mention.
        </p>
        <p>
          Syntactic features Syntactic structure can shed a lot of light on the NPs’
discourse roles. Studies in Centering theory and various discourse studies showed
that coreferent mentions tend to be core verbal arguments and prefer
sentenceinitial positions in a sentence (e.g. [
          <xref ref-type="bibr" rid="ref4">5,19</xref>
          ]).
        </p>
        <p>In order to include this information in the classifier we used the syntactic
annotation described in section 3.1. An NP was considered a subject if the type
of a syntactic role was ‘предик’. An NP was considered an object if its role was
‘1-компл’4.
4 A submitted version of a paper did not use syntactic information because the
experiments showed its low quality. Further experiments showed that although there
are a lot of errors in the syntactic information, it still improves the classification
quality.</p>
        <p>Less standard feature that we employed was if an NP is in genitive case. The
source for this feature was the intuition about Russian genitive that it coincides
with non-argument positions.</p>
        <p>Lexical features We used four precompiled lists to detect non-coreferent mentions:
(i) indefinites, (ii) possessive pronouns, (iii) negative pronouns and (iv) non–
identity words (‘takoj zhe’ such as, ‘similar’ podobnyj, etc.). If all the previous
groups contained features that were designed to detect NPs that are more likely
to be unique in the text, judging by their discourse role, those lists (except for
the last one) should detect NPs that can not be coreferent. This way we have
features for both non-referring and potentially coreferent but unique mentions.
3.3</p>
      </sec>
      <sec id="sec-3-3">
        <title>Results</title>
        <p>
          To test how good various features distinguish singleton mentions from
nonsingleton ones, we have built a set of classifiers using Random Forest classifier
implemented in Scikit-Learn library ([
          <xref ref-type="bibr" rid="ref9">10</xref>
          ])5. This classifier was chosen mainly
for two reasons: firstly, Decision Tree classifiers and improvements over them
are often used in related work. Secondly, using this algorithm provided some
insight on feature importances during feature engineering step. As a baseline,
we established a heuristic baseline: an NP is considered a singleton mention if
and only if there were no such NPs or no NPs with the same head before. Results
of the experiments are presented in table 1.
        </p>
        <p>
          As it has already been noted, singletons are more frequent than coreferent
mentions. Due to this disproportion the training set is unbalanced there
are more than two times more singleton mentions than non-singleton ones
and this influences resulting classification quality. To overcome this problem, we
performed oversampling on the training set. The best results were achieved using
SMOTE+Tomek method ([
          <xref ref-type="bibr" rid="ref2">3</xref>
          ]). The results are presented in table 1
P R F1
Baseline 0.470 0.665 0.551
Basic 0.621 0.630 0.626
Basic + Struct 0.601 0.676 0.637
Basic + Struct + Lists 0.609 0.671 0.638
        </p>
        <p>All features 0.600 0.708 0.650
Таблица 1. Classification results (for the minority class)</p>
        <p>The basic feature set gives the highest precision, while the recall is lower
than using more complex features. The more features are used, the more is
recall. Using the full set of features increases the recall by about 12%.
5 An IPython notebook with the experiments is available here: https://git.io/vwE9F.</p>
        <p>To understand the individual importance of each feature, we have trained a
logistic regression model on the training data. The coefficients for each feature
are given in table 2. A singleton is the positive class, so positive coefficients mean
a bias towards being a singleton mention.</p>
        <p>Estimate</p>
        <p>
          Results for basic features agree with the results reported in [
          <xref ref-type="bibr" rid="ref12">13</xref>
          ] for English.
The unique head lexeme is a sustainable feature for the unique mentions. Animate
and proper nouns, and anaphoric pronouns are more likely to be coreferential.
        </p>
        <p>
          The NP length (number of modifiers) acts in a less expected way: short and
moderately long NPs tend to be non-singletons whereas only very long NPs tend
to be singletons. However, this agrees with the result in [
          <xref ref-type="bibr" rid="ref12">13</xref>
          ]. Positive number of
adjectives correlates with being coreferent as expected.
        </p>
        <p>The syntactic features also have expected behavior: subject mentions tend
to be non-singletons and mentions in genitive are slightly biased towards being
singletons. This agrees with previously stated intuition of them being non-topical
arguments.</p>
        <p>Indefinite (free-choice) pronouns such as lyuboj ‘any’, kazhdyj ‘every’,
ktonibud ‘anybody’ tend to be used within unique mentions in the discourse. The
most reliable feature for different types of pronouns is the negative pronoun type.
This result goes in hand with theoretical assumptions that free-choice pronouns
and negative pronouns are non-referential expressions and, thus, they are less
probable in in a coreferential chain.
4</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Discussion</title>
      <p>The first experiment on the singleton detection has shown that the basic set
of features gives relatively high precision and a low recall. Additional features
(those taken from previous research for English as well as, relevant for
firstmention detection) improve the recall, though there is a small precision loss. As
a result, at least 70% of singletons can be filtered out from the mentions set for
coreference pairs generation in coreference resolution task which will drastically
improve computational efficiency and should improve the resolution quality.
5</p>
    </sec>
    <sec id="sec-5">
      <title>Conclusions</title>
      <p>In this work we discussed various features used for the singleton detection task as
a subtask of coreference resolution systems. We presented a preliminary research
on singleton detection with a special emphasis on Russian.</p>
      <p>The focus was on the testing features introduced for the corresponding task
for English in the previous work, as well as some features used for the
discoursenew detection.</p>
      <p>We tested classifiers that can distinguish singletons. We also set a baseline
for further experiments and tested special lexical features. Though the recall is
low we think that the results can be used in the coreference resolution systems
to filter out false candidates for the coreferring mentions.</p>
      <p>The analysis of the results has shown that some of the lexical features such
as special class of pronoun types, for example, negative and free choice pronouns
are quite promising for this task and need further investigation and enhancing.
The other promising direction is a more detailed investigation of the impact
of the grammatical role of a given NP. In our future work we are planning to
examine the contribution of more elaborated features.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>The authors would like to thank the Lomonosov Moscow University students
who participated in the corpus markup, Dmitrij Gorshkov for creating and
maintaining corpus annotation software, and Dmitrij Privoznov for his useful
comments.</p>
      <p>This research was supported by the grant from Russian Foundation for Basic
Research Fund (15-07-09306).
Список литературы
1. Ariel, M.: Accessing Noun-Phrase Antecedents. Routledge (1990)
18. Uryupina, O.: High-precision identification of discourse new and unique noun
phrases. In: ACL Student Workshop. Sapporo (2003)
19. Ward, G., Birner, B.: Information structure and non-canonical syntax. The
handbook of pragmatics pp. 153–174 (2004)</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          2.
          <string-name>
            <surname>Arutyunova</surname>
          </string-name>
          , N.: Nomination, reference, meaning. [nominaciya, referenciya, znacheniye]
          <article-title>(in Russian)</article-title>
          . In: Nomination: General Questions. [Nominaciya: obshie voprosi].
          <source>Nauka</source>
          (
          <year>1980</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          3.
          <string-name>
            <surname>Batista</surname>
            ,
            <given-names>G.E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bazzan</surname>
            ,
            <given-names>A.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Monard</surname>
            ,
            <given-names>M.C.</given-names>
          </string-name>
          :
          <article-title>Balancing training data for automated annotation of keywords: a case study</article-title>
          .
          <source>In: WOB</source>
          . pp.
          <fpage>10</fpage>
          -
          <lpage>18</lpage>
          (
          <year>2003</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          4.
          <string-name>
            <surname>Bonch-Osmolovskaya</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Toldova</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Klintsov</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          :
          <article-title>Introductory noun phrases: a case of mass media texts. [strategii introduktivnoj nominacii v teksrah smi] (in Russian) (</article-title>
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          5.
          <string-name>
            <surname>Grosz</surname>
            ,
            <given-names>B.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weinstein</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joshi</surname>
            ,
            <given-names>A.K.</given-names>
          </string-name>
          :
          <article-title>Centering: A framework for modeling the local coherence of discourse</article-title>
          .
          <source>Comput. Linguist</source>
          .
          <volume>21</volume>
          (
          <issue>2</issue>
          ),
          <fpage>203</fpage>
          -
          <lpage>225</lpage>
          (
          <year>Jun 1995</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          6.
          <string-name>
            <surname>Hirschman</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chinchor</surname>
          </string-name>
          , N.:
          <article-title>Appendix f: Muc-7 coreference task definition (version 3.0)</article-title>
          .
          <source>In: Seventh Message Understanding Conference (MUC-7): Proceedings of a Conference Held in Fairfax, Virginia, April 29 - May 1</source>
          ,
          <year>1998</year>
          (
          <year>1998</year>
          ), http://www.aclweb.org/anthology/M98-1029
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          7.
          <string-name>
            <surname>Ionov</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kutuzov</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>Influence of morphology processing quality on automated anaphora resolution for russian</article-title>
          .
          <source>In: Proceedings of the international conference Dialogue-2014</source>
          . RGGU (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          8.
          <string-name>
            <surname>Mitkov</surname>
          </string-name>
          , R.:
          <article-title>Anaphora resolution: the state of the art (</article-title>
          <year>1999</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          9.
          <string-name>
            <surname>Ng</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cardie</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Identifying anaphoric and non-anaphoric noun phrases to improve coreference resolution</article-title>
          .
          <source>In: Proceedings of the 19th international conference on Computational linguistics-Volume</source>
          <volume>1</volume>
          . pp.
          <fpage>1</fpage>
          -
          <lpage>7</lpage>
          . Association for Computational Linguistics (
          <year>2002</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          10.
          <string-name>
            <surname>Pedregosa</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Varoquaux</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gramfort</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Michel</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Thirion</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grisel</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Blondel</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Prettenhofer</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Weiss</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dubourg</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vanderplas</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Passos</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cournapeau</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Brucher</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Perrot</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Duchesnay</surname>
          </string-name>
          , E.:
          <article-title>Scikit-learn: Machine learning in Python</article-title>
          .
          <source>Journal of Machine Learning Research</source>
          <volume>12</volume>
          ,
          <fpage>2825</fpage>
          -
          <lpage>2830</lpage>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          11.
          <string-name>
            <surname>Poesio</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ponzetto</surname>
            ,
            <given-names>S.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Versley</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          :
          <article-title>Computational models of anaphora resolution: A survey (</article-title>
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          12.
          <string-name>
            <surname>Prince</surname>
            ,
            <given-names>E.F.</given-names>
          </string-name>
          :
          <article-title>The zpg letter: Subjects, definiteness, and information-status. Discourse description: diverse analyses of a fund raising text pp</article-title>
          .
          <fpage>295</fpage>
          -
          <lpage>325</lpage>
          (
          <year>1992</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          13.
          <string-name>
            <surname>Recasens</surname>
          </string-name>
          , M.,
          <string-name>
            <surname>de Marneffe</surname>
            ,
            <given-names>M.C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Potts</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>The life and death of discourse entities: Identifying singleton mentions</article-title>
          . In: Human Language Technologies:
          <article-title>The 2013 Annual Conference of the North American Chapter of the Association for Computational Linguistics</article-title>
          . pp.
          <fpage>627</fpage>
          -
          <lpage>633</lpage>
          . Association for Computational Linguistics, Stroudsburg, PA (
          <year>June 2013</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          14.
          <string-name>
            <surname>Sharoff</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nivre</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>The proper place of men and machines in language technology: Processing Russian without any linguistic knowledge</article-title>
          .
          <source>In: Proc. Dialogue</source>
          , Russian International Conference on Computational Linguistics.
          <source>Bekasovo</source>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          15.
          <string-name>
            <surname>Toldova</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          :
          <article-title>Struktura diskursa i mehanizm fokusirovaniya kak vazhnie faktori vibora nominatsii ob'ekta v tekste (Discourse structure and the focusing mechanism as important factors of referential choice in text</article-title>
          ) (
          <year>1994</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          16.
          <string-name>
            <surname>Toldova</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ionov</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Features for discourse-new referent detection in russian</article-title>
          .
          <source>In: Computational Linguistics and Intelligent Text Processing - 17th International Conference, CICLing</source>
          <year>2016</year>
          , Konya, Turkey, April 3-
          <issue>9</issue>
          ,
          <year>2016</year>
          , Proceedings. Konya, Turkey (in press)
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          17.
          <string-name>
            <surname>Toldova</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rojtberg</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ladygina</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vasilyeva</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Azerkovich</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kurzukov</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ivanova</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nedoluzhko</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Grishina</surname>
          </string-name>
          , J.:
          <article-title>RU-EVAL-2014: Evaluating Anaphora and Coreference Resolution for Russian</article-title>
          .
          <source>Computational Linguistics and Intellectual Technologies</source>
          <volume>13</volume>
          (
          <issue>20</issue>
          ),
          <fpage>681</fpage>
          -
          <lpage>694</lpage>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>