<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Coreference Resolution Algorithm for Ukrainian-Language Texts Using Decision Trees</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Sergiy Pogorilyy</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Pavlo Biletskyi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Taras Shevchenko National University of Kyiv</institution>
          ,
          <addr-line>60 Volodymyrska Street, Kyiv, 01033</addr-line>
          ,
          <country country="UA">Ukraine</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The paper examines the problem of coreference resolution in Ukrainian-language texts using decision trees. An application that uses vector representations of Elmo words and other characteristics for the automated formation of a decision tree has been developed. A set of prepared texts containing more than 360,000 words was used to form the decision tree and evaluate the accuracy of the algorithm. The decision tree created to determine whether a pair of objects is coreference was used to form clusters of coreference objects. Special metrics were used for comparison with the results obtained by other algorithms in the Ukrainian language. Coreference resolution, natural language processing (NLP), decision trees, artificial intelligence (AI), vector words representation, neural networks.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>context [1].
underlined).</p>
      <p>Natural language processing (NLP) is a large field that includes many tasks: translation between
natural languages, human-computer interfaces, analysis and generation of natural speech, information
extraction. One of the tasks of natural language processing is coreference resolution. Coreferentiality
in texts means a relationship between syntactic units indicating the same object (referent) in a given
Below are examples of coreference (the referent is highlighted in bold, the pronouns are
Simple anaphora example:
Simple cataphora example:</p>
      <p>He crossed the mountain. It was high.</p>
      <p>An example of a compound antecedent:</p>
      <p>She took the road to the right. Mery was in a good mood today.</p>
      <p>John, David and Julia were tired. They all worked underground.
Ukrainian language has an arbitrary word order, just like other Slavic languages - Polish, Russian,</p>
      <p>Karpo prykynuv take slivtse, shcho batko perestav struhaty i pochav pryslukhatys. Vin hlianuv na
syniv cherez khvorostianu stinu. Syny stoialy bez dila y balakaly, pospyravshys na zastupy (Ivan
Nechui-Levytskyi, «Kaidasheva simia», original word order).</p>
      <p>From the second sentence in the example, it is possible to rearrange the words to form other
grammatically correct sentences with very close meanings, but with different accents, depending on
the author's intention:
•</p>
      <p>Cherez khvorostianu stinu vin hlianuv na syniv.</p>
      <p>2022 Copyright for this paper by its authors.
• Na syniv vin hlianuv cherez khvorostianu stinu.
• Cherez stinu khvorostianu vin hlianuv na syniv.
• Cherez khvorostianu stinu hlianuv vin na syniv.
• Cherez khvorostianu stinu na syniv hlianuv vin.</p>
      <p>At the same time, only one combination is possible in English: «He looked at his sons through the
twig wall», because it uses the standard subject-verb-object (SVO, subject verb object) word order. In
other languages, the subject-object-verb (SOV, subject object verb) order is also common. In this
regard, the algorithm for the Ukrainian language should be able to work with different word orders.</p>
      <p>Coreference resolution allows finding connections between sentences and within them, extracting
information from texts, improving the results of text analysis in other tasks, such as translation from
one language to another, assessment of the coherence of texts.</p>
      <p>At the initial stages of research, algorithms based on rules manually formed by experienced
linguists were used to search for coreference objects. Such algorithms were c reated for a specific
language and had to take into account many features to achieve high results.</p>
      <p>Over time, automated approaches such as neural networks and decision trees began to be used to
solve the problem. They do not require manual rule creation, but require large data sets to prepare
them. Nevertheless, often, automated algorithms use simple rules to form initial clusters.</p>
      <p>In previous works, such as [2], [3], usage of the decision trees was considered for coreference
resolution and utilized on MUC-6 dataset [2] for English language and CoNLL-2012 [3] for Arabic,
Chinese and English, but decision trees appliance was newer tested on Ukrainian -language datasets.</p>
      <p>The article proposes a method for coreference resolution in Ukrainian language using decision
trees. An application has been developed that uses Elmo vector representations of words and other
features for automated decision tree formation. A set of prepared texts containing more than 360,000
words was used to form the tree and evaluate the accu racy of the algorithm. The tree formation
parameters are selected. The decision tree created to determine whether a pair of objects is
coreference used to form clusters of coreference objects. Special metrics were used for comparison
with the results obtained by other algorithms in the Ukrainian language.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Means used to implement the coreference resolution algorithm</title>
      <p>The application uses vector representations of words obtained by applying the ELMo library[ 4].
This library allows converting words into vectors corresponding to their semantic, lexical, and
syntactic meaning. Unlike other libraries that allow to form vector representations of words, such as
Word2Vec[5], vector representations formed by ELMo take into account not only the meaning of a
single word, but also the meaning of surrounding words, which allows to better find connections
between individual words and sentences. ELMo uses neural networks to obtain vector representations
of words and requires training. Version used in application adapted for the Ukrainian language.</p>
      <p>The Scikit-learn library [6] is used. This library includes many tools for solving regression,
clustering and classification problems, in particular, it contains an optimized implementation of
decision trees with the ability to conf igure tree construction parameters.</p>
      <p>One of the main advantages of decision trees over other algorithms is the ability to visualize the
constructed tree. This allows conducting analysis of created tree and it`s internal operational logics.
Also, decisions automatically made by the tree on each of its steps allow to improve the analysis of
data, find dependencies between parameters, determine the limit of overfitting and adjust parameters
for forming decision tree. Furthermore, after analysis completed, it is possible to make changes in
decision tree structure and see, if changes are improving the result. For this, the Graphviz library [7],
which is specifically developed for graph visualization was used.</p>
      <p>A prepared corpus of Ukrainian-language texts containing more than 360,000 words (&gt; 2,500
texts) was used to create a decision tree. The marking of coreference objects in it is carried out
manually, and for obtaining additional information (gender, number, lemmatized (initial) version of
the word) the UDpipe library [8], which uses neural networks and is trained on the Ukrainian text
corpus is used.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Data format for representing coreference objects and their analysis</title>
      <p>To analyze texts using decision trees, they need to be presented in the correct form. The follow ing
characteristics are used in the work to describe coreference objects (considered in the article [ 8]):
• Cosine similarity of vectors of the semantic representation of objects under consideration. For
this, a pre-trained ELMo model was used. In the work, if the number of words included in the
object &gt; 1, the arithmetic mean of word vectors is used.
• The number of words between the selected objects.
• The number of objects between the selected potentially coreference objects.
• Boolean, true if the first object is a pronoun.
• Boolean, true if the second object is a pronoun.
• Boolean, true if the lemmatized versions (initial word forms) of the objects match. In the
algorithm, the matching of lemmatized versions is defined as the matching of at least one word in
both objects under consideration.
• Boolean, true if both objects have the same number (singular or plural).
• Boolean, true if both objects have the same genus.
• Boolean, true if both objects are proper names.</p>
      <p>The input of the algorithm is a text consisting of ind ividual words, punctuation, and additional
information prepared using the UDpipe library. This includes gender, number, lemmatized version of
the word, part of speech. Also, for words included in coreference groups, an identifier is added, which
allows to assign a specific word or word combination to a coreference group (prepared manually).</p>
      <p>
        For the algorithm operation, a Python list is formed for each text under consideration, containing
the indexes of the words included in the word combinations, which are potentially coreferential
objects, before combining them into coreferential clusters, but in a format that facilitates their
subsequent union (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ), – is a separate word, - is a word combination.
(
        <xref ref-type="bibr" rid="ref1">1</xref>
        )
(
        <xref ref-type="bibr" rid="ref2">2</xref>
        )
(
        <xref ref-type="bibr" rid="ref3">3</xref>
        )
(
        <xref ref-type="bibr" rid="ref4">4</xref>
        )
      </p>
      <p>
        Also, lists containing correctly formed clusters of coreference objects are created. These lists are
necessary at the stage of comparing the clusters obtained as a result of predictions of the decision tree
and valid clusters (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ),
,
- is a separate cluster.
      </p>
      <p>
        Since the clustering task is reduced in the algorithm to a classification task, to create decision trees
and predictions with their help, lists with the parameters of each pair of potentially coreferential
objects (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ) are needed, which include cosSim – the cosine similarity of objects, nWBtw – the number
of words between objects under consideration, nObjBetw – the number of objects between the objects
under consideration, len1 – the length (number of words) of the first object, len2 – the length of the
second object, 1pron – whether the first object is a pronoun, 2pron – whether the second object is a
pronoun, 1prp – whether the first object is a proper name, 2prp – whether the second object is a
proper name, lemS – whether the lemmatized versions of the objects match, gendS – whether the
objects have the same gender , numS– whether the objects have the same number.
      </p>
      <p>
        Each pair of objects is denoted as x (
        <xref ref-type="bibr" rid="ref4">4</xref>
        ).
      </p>
      <p>
        Labels indicating whether the pair of objects under consideration are coreferential are also
required to check algorithm (
        <xref ref-type="bibr" rid="ref5">5</xref>
        ).
      </p>
      <p>
        After creating a decision tree, when using it for predictions from list (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ), a list (
        <xref ref-type="bibr" rid="ref5">5</xref>
        ) containing a list of
predicted clusters is generated.
(
        <xref ref-type="bibr" rid="ref5">5</xref>
        )
(
        <xref ref-type="bibr" rid="ref6">6</xref>
        )
      </p>
      <p>
        The resulting lists with characteristics of groups of coreference objects (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ) as well as their
labeling (
        <xref ref-type="bibr" rid="ref4">4</xref>
        ), containing 2,400,000 samples of coreference and non -coreference objects, are divided
into two parts - the first (1500 texts, ~ 60%) is used to form a decision tree, the second (1015 texts, ~
40%) – to check the effectiveness of the algorithm (analysis of the obtained results).
      </p>
    </sec>
    <sec id="sec-4">
      <title>4. Algorithm for coreference resolution using decision trees</title>
    </sec>
    <sec id="sec-5">
      <title>4.1. Formation of the decision tree</title>
      <p>The decision tree is implemented using the sklearn library, which has the decision tree class
sklearn.tree.DecisionTreeClassifier. When creating an instance of a class, the parameters of the
decision tree are determined - the criteria for splitting into subtrees during formation, the maximum
depth of the tree, the minimum number of elements for splitting, the minimum number of elements in
one leaf of the tree, weights for the expected classes, and others. By default, the parameters of the tree
are specified to obtain an error-free configuration of the tree on the sample used for its formation.
This approach leads to excessive adaptation of the tree to the data used in its formation and reduces
the accuracy of work on sets that were not p reviously analyzed by the algorithm. Also, for large
volumes of data, the use of such a configuration creates an excessively large tree, the formation of
which takes a long time.</p>
      <p>Thus, it becomes necessary to limit the size of the tree in order to achieve higher results on the data
that were not used to form the tree. For this, the parameter min_impurity_decrease was used, which
allows determining the minimally sufficient value of reducing heterogeneity in subsequent subtrees
during splitting. Unlike other ways of limiting the size of the tree, such as limiting the depth or
limiting the number of elements in the leaves, this indicator allows more evenly limit the size of the
tree. For the formation of final version of the decision tree, parameter of min_impu rity_decrease was
set to 0.000003.</p>
      <p>
        To form a tree, prepared data of pairs of potentially coreferential objects (
        <xref ref-type="bibr" rid="ref3">3</xref>
        ) with a marking of
whether they are coreferential (
        <xref ref-type="bibr" rid="ref4">4</xref>
        ) are submitted to the fit function. The output is a tree capable of
analyzing pairs of coreference objects. Thus, the problem of coreference object clustering is reduced
to the problem of decision tree classification.
      </p>
      <p>Part of the resulting tree is shown in Fig. 1. For illustration purposes, the depth of the tree is
artificially limited to fit the resulting tree on the screen. As can be seen from the figure, the division
into coreferential and non-coreferential objects begins with the characteristic that allows the best
separation of the current group of objects - the coincidence of lemmatized versions of objects.
Further, thanks to the use of decision trees, the following logic can be followed in the subtree of the
tree (Fig. 2) – if the lemmatized versions of the objects match, the length of the first and second
objects is 1, and the first object is a proper name, then with a high probability a pair of objects is
coreferential.</p>
      <p>Further filtering allows finding more coreferent objects: in Fig. 3 if first object is not proper
noun but pronoun, then it is considered coreferent to the second object. On Fig. 4 if first object is
proper noun and its length is 2, then it is also considered coreferent.</p>
      <p>Decision trees are also allow filtering of negative coreference examples, as it is shown at Fig.
5, on which there is subtree of the decision tree right after initial filtering if lemmatized versions of
objects are not same. If second object is not pronoun and cosine similarity measure of objects is lower
then 0.426, then with high probability objects are not coreferent.</p>
    </sec>
    <sec id="sec-6">
      <title>Using a decision tree for coreference resolution</title>
      <p>
        After forming a decision tree, the created tree used to obtain clusters of coreference objects: pairs
of objects - candidates for coreference are considered (
        <xref ref-type="bibr" rid="ref1">1</xref>
        ). For each pair of objects, the parameters (
        <xref ref-type="bibr" rid="ref3">3</xref>
        )
necessary for the classification of objects by a decision tree are determined. If th e pair is classified to
be coreferential, the objects are merged. After the formation of clusters containing several objects,
their merging occurs if at least one pair of objects from the first and second clusters is recognized to
be coreferential. During the experimental studies, it was found that when several cycles of passes are
used, while cluster merging is possible, the results on the metrics (next paragraph) were 2 -5% higher
than without the use of cyclic passes, therefore, in the final version of th e algorithm, several passes
are used, while during cycle there is at least one merging. As a result of the algorithm, a list of
clusters containing coreference objects within common clusters is formed (
        <xref ref-type="bibr" rid="ref5">5</xref>
        ).
      </p>
    </sec>
    <sec id="sec-7">
      <title>Analysis of the obtained results</title>
      <p>
        The metrics [9] and MUC [10] are used to evaluate the results obtained while using the
coreference resolution algorithm. These metrics allow comparing groups of clusters - with the correct
ordering of coreference objects (
        <xref ref-type="bibr" rid="ref2">2</xref>
        ) and with the predicte d ordering (
        <xref ref-type="bibr" rid="ref5">5</xref>
        ), numerically reflecting the
difference between clusters. For each metric, the precision (ratio of correctly selected objects to all
selected objects), recall (ratio of correctly selected objects to all objects belonging to this cluster), and
F1 measure (mean harmonic of precision and recall) shown.
      </p>
      <p>The metric is used in a wide range of clustering problems. The metric considers individual
elements in the list of predicted clusters, for which the integral value is calculated. The precision for
the metric is defined as the arithmetic average of the accuracy for each element:
where – number of elements in the true coreference cluster, - number of subgroups
into which the true coreference cluster is divided by the assumed clusters. Precision is defined as:
where n - number of the selected element, N – number of elements in the list, - number of
elements that belong to the same coreference group as the selected element (including the selected
element) and are included in the predicted cluster, - total number of elements in the predicted
cluster.</p>
      <p>Recall for metric is defined as the arithmetic mean of recall for each element:
where n - number of the selected element, N - number of elements in the list, - number of
elements that belong to the same coreference group as the selected element (including the selected
element) and are included in the expected cluster, - total number of elements in the same group as
the selected element.</p>
      <p>
        The MUC metric is specially designed to evaluate the performance of algorithms that solve
coreference resolution task. The MUC metric considers the entire list of clusters for which indications
are calculated. For MUC, recall is determined by the formula:
(
        <xref ref-type="bibr" rid="ref7">7</xref>
        )
(
        <xref ref-type="bibr" rid="ref8">8</xref>
        )
(
        <xref ref-type="bibr" rid="ref9">9</xref>
        )
where – the number of elements in the assumed coreference cluster, - the number
of subgroups into which the true clusters divid e by the assumed coreference cluster.
      </p>
      <p>For each metric, the F1 measure is calculated, determined by the formula:
where R - recall, P - precision.</p>
      <p>The evaluation of results of the obtained decision tree is performed on the part of the corpus
that was not used during trees formation (1015 texts). During experimental research, the optimal value
of the min_impurity_decrease parameter was determined, for which the highest results on the and
MUC metrics were achieved. The results of a comparison of decision trees algorithm with other
algorithms used for the analysis of Ukrainian -language texts [1, 9] are shown in the table. 1.</p>
      <p>For the metric , the decision tree shows highest results compared to other approaches. On the
MUC metric, the decision tree approach shows significantly higher results than those achieved using
the convolutional neural network (CNN) [9] and the natural language model RoBERTa [1], similar to
a two-way neural network with long and short-term memory (BiLSTM) [9]. It is worth noting that the
algorithm's precision for this decision tree on the MUC and metrics is the highest compared to
other approaches, while the completeness similar to the BiLSTM variant with one pass. Parameters of
the decision tree formation allow to increase its precision by reducing the recall or vice versa, thus
adapting to cases, for which accuracy for found coreference objects (precision) or finding the majority
of coreference objects (recall) is more important.</p>
    </sec>
    <sec id="sec-8">
      <title>6. Conclusions</title>
      <p>In the work, the task of coreference resolution for Ukrainian language is considered. An
application that uses decision trees to search for coreference objects is created: data was prepared for
analysis, a decision tree was built, its parameters were modified to achieve higher results, the
performance of the application was evaluated on and MUC metrics. Decision trees, thanks to the
possibility of graphical representation, make it easier to analyze the obtained results and explain how
one or another result is obtained, which distinguishes them from other popular methods of data
analysis, in particular neural networks.</p>
      <p>The obtained results show a relatively high efficiency of the algorithm - the highest results were
achieved on the metric and relatively high on the BiLSTM metric, which indicates the expediency
of using decision trees to search for coreference objects in Ukrainian -language texts.</p>
    </sec>
    <sec id="sec-9">
      <title>7. References</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Pogorilyy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Biletskyi</surname>
          </string-name>
          ,
          <article-title>Usage of a graphics processor to accelerate coreference resolution while using the RoBERTa model</article-title>
          .,
          <source>Scientific works of DonNTU</source>
          , Series: Informatics, cybernetics and computer technology,
          <year>2022</year>
          , pp.
          <fpage>4</fpage>
          -
          <lpage>9</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Dzunic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Momcilovic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Todorovic</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Stankovic</surname>
          </string-name>
          ,
          <article-title>Coreference Resolution Using Decision Trees, 2006 8th Seminar on Neural Network Applications in Electrical Engineering</article-title>
          ,
          <year>2006</year>
          , pp.
          <fpage>109</fpage>
          -
          <lpage>114</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>E.</given-names>
            <surname>Fernandes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Nogueira</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Milidiú</surname>
          </string-name>
          .
          <article-title>Latent Trees for Coreference Resolution</article-title>
          .
          <source>Computational Linguistics</source>
          <volume>40</volume>
          (
          <issue>4</issue>
          ),
          <year>2014</year>
          , pp.
          <fpage>801</fpage>
          -
          <lpage>835</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M.</given-names>
            <surname>Peters</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Neumann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Iyyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gardner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Clark</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Zettelemoyer</surname>
          </string-name>
          ,
          <article-title>Deep contextualized word representations</article-title>
          .
          <source>In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Hum an Language Technologies</source>
          ,
          <year>2018</year>
          , pp.
          <fpage>2227</fpage>
          -
          <lpage>2237</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>T.</given-names>
            <surname>Mikolov</surname>
          </string-name>
          , I. Sutskever,
          <string-name>
            <given-names>K.</given-names>
            <surname>Chen</surname>
          </string-name>
          , G. Corrado,
          <string-name>
            <given-names>J.</given-names>
            <surname>Dean</surname>
          </string-name>
          ,
          <article-title>Distributed Representations of Words and Phrases and their Compositionality</article-title>
          .
          <source>Proceedings of the 26th International Conference on Neural Information Processing Systems</source>
          ,
          <year>2013</year>
          , pp.
          <fpage>3111</fpage>
          -
          <lpage>3119</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          <article-title>[6] Scikit-learn library</article-title>
          . URL: https://scikit-learn.org/ .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          <article-title>[7] Graphviz library</article-title>
          . URL: https://graphviz.org/ .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          <article-title>[8] UDpipe library</article-title>
          . URL: https://lindat.mff.cuni.cz/services/udpipe/ .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>S.</given-names>
            <surname>Telenyk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Pogorilyy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kramov</surname>
          </string-name>
          ,
          <article-title>The complex method of coreferent clusters detection based on a BiLSTM neural network</article-title>
          ,
          <source>Knowledge Based Systems</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>205</fpage>
          -
          <lpage>210</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bagga</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Baldwin</surname>
          </string-name>
          , Algorithms for Scoring Coreference Chains,
          <source>The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference</source>
          ,
          <year>1998</year>
          , pp.
          <fpage>563</fpage>
          -
          <lpage>566</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>M.</given-names>
            <surname>Vilain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Burger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Aberdeen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Connolly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Hirschman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A</given-names>
            <surname>Model-Theoretic Coreference Scoring Scheme</surname>
          </string-name>
          ,
          <source>Proceedings of the 6th Conference on Message Understanding (MUC)</source>
          ,
          <year>1995</year>
          , pp.
          <fpage>45</fpage>
          -
          <lpage>52</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>