<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Annotated su x tree similarity measure for text summarization</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>National Research University - Higher School of Economics Moscow</institution>
          ,
          <country country="RU">Russia</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>The paper describes an attempt to improve the TextRank algorithm. TextRank is an algorithm for unsupervised text summarisation. It has two main stages: first stage is representing a text as a weighted directed graph, where nodes stand for single sentences, and edges are weighted with sentence similarity and connect sequential sentences. The second stage is applying the PageRank algorithm [1] as is to the graph. The nodes that get the highest ranks form the summary of the text. We focus on the first stage, especially on measuring the sentence similarity. Mihalcea and Tarau [4] suggest to employ the common scheme: use the Vector space model (VSM), so that every text is a vector in the space of words or stems, and compute cosine similarity between these vectors. Our idea is to replace this scheme by using the annotated su x trees (AST) [5] model for sentence representation. The AST overcomes several limitations of the VSM model, such as being dependent on the size of vocabulary, the length of sentences and demanding stemming or lemmatisation. This is achieved by taking all fuzzy matches between sentences into account and computing probabilities of matched coocurrencies. More specifically we develop an algorithm for common subtree construction and annotation. The common subtrees are used to score the similarity between two sentences. Using this algorithm allows us to achieve slight improvements according to cosine baseline on our own collection of Russian newspaper texts. The AST measure gained around 0.05 points of precision more than the cosine measure. This is a great figure for natural language processing task, taking into account how low the baseline precision of the cosine measure is. The fact that the precision is so low can be explained by some lack of consistency in the constructed collection: the authors of the articles use di↵ erent strategies to highlight the important sentences. The text collection is heterogeneous: in some articles there are 10 or more sentences highlighted, in some only the first one. Unfortunately, there is no other test collection for text summarisation in Russian. For further experiments we might need to exclude some articles, so that the size of summary would be more stable. Another issue of our test collection is the selection of sentences that form summaries. When the test collections are constructed manually, summaries are chosen to common principles. But we can not be sure that the sentences are not highlighted randomly. Although the AST technique is rather slow, it is not a big issue for the text summarisation problem. The summarisation problem is not that In: P. Cellier, T. Charnois, A. Hotho, S. Matwin, M.-F. Moens, Y. Toussaint (Eds.): Proceedings of DMNLP, Workshop at ECML/PKDD, Porto, Portugal, 2015. Copyright c by the paper's authors. Copying only for private and academic purposes.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>
        M. Yakovlev and E. Chernyak
kind of problems where on-line algorithms are required. Hence the
precision plays more significant part than time characteristics.
There are several directions of future work. First of all, we have to
conduct experiments on the standard DUC (Document Understanding
Conference [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]) collections in English. Second, we are going to develop
different methods for construction and scoring of common subtrees and
compare it to each other. Finally, we may use some external and more
e cient implementation of the AST method, such as EAST Python
library by Mikhail Dubov [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], which uses annotated su x arrays. More
details on this work can be found in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>Keywords: TextRank, annotated su x tree</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Brin</surname>
            <given-names>S.</given-names>
          </string-name>
          , Page L. :
          <article-title>The anatomy of a large-scale hypertextual Web search engine</article-title>
          .
          <source>In: Proceedings of the seventh international conference on World Wide Web</source>
          <volume>7</volume>
          ,
          <fpage>107</fpage>
          -
          <lpage>117</lpage>
          (
          <year>1998</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>2. Document Understanding Conference, http://www-nlpir.nist.gov/</mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>Enhanced</given-names>
            <surname>Annotated Su x Tree</surname>
          </string-name>
          , https://pypi.python.org/pypi/EAST/0.2.2/
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Mihalcea</surname>
            <given-names>R.</given-names>
          </string-name>
          , Tarau P. :
          <article-title>TextRank: bringing order into text</article-title>
          .
          <source>In: Proceedings of the Conference on Empirical Methods in Natural Language Processing</source>
          ,
          <fpage>404</fpage>
          -
          <lpage>411</lpage>
          (
          <year>2004</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Pampapathi</surname>
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mirkin</surname>
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Levene</surname>
            <given-names>M.</given-names>
          </string-name>
          (
          <year>2008</year>
          )
          <article-title>: A su x tree approach to anti-spam email filtering</article-title>
          .
          <source>In: Machine Learning</source>
          ,
          <volume>65</volume>
          (
          <issue>1</issue>
          ),
          <fpage>309</fpage>
          -
          <lpage>338</lpage>
          (
          <year>2008</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Yakovlev</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chernyak</surname>
            <given-names>E.</given-names>
          </string-name>
          (
          <year>2015</year>
          )
          <article-title>: Using annotated su x tree similarity measure for text summarization</article-title>
          .
          <source>In: Proccedings of European Conferenece on Data Analysis</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>