<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Dense Semantic Graph and its Application in Single Document Summarisation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Monika Joshi</string-name>
          <email>joshi-m@email.ulster.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Hui Wang</string-name>
          <email>H.Wang@ulster.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sally McClean</string-name>
          <email>sally@infc.ulst.ac.uk</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>University of Ulster</institution>
          ,
          <addr-line>Co. Antrim, BT37 0QB</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Ulster</institution>
          ,
          <addr-line>Co. Londonderry, BT52 1SA</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Semantic graph representation of text is an important part of natural language processing applications such as text summarisation. We have studied two ways of constructing the semantic graph of a document from dependency parsing of its sentences. The first graph is derived from the subject-object-verb representation of sentence, and the second graph is derived from considering more dependency relations in the sentence by a shortest distance dependency path calculation, resulting in a dense semantic graph. We have shown through experiments that dense semantic graphs gives better performance in semantic graph based unsupervised extractive text summarisation.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>Information can be categorized into many forms -- numerical, visual, text, and audio.
Text is abundantly present in online resources. Online blogs, Wikipedia knowledge
base, patent documents and customer reviews are potential information sources for
different user requirements. One of these requirements is to present a short summary
of the originally larger document. The summary is expected to include important
information from the original text documents. This is usually achieved by keeping the
informative parts of the document and reducing repetitive information.</p>
      <p>There are two types of text summarization: multiple document summarisation and
single document summarization. The former is aimed at removing repetitive content
in a collection of documents. The latter is aimed at shortening a single document
whilst keeping the important information. Single document summarisation is
particularly useful because large documents are common especially in the digital age, and
shortening them without losing important information is certain to save time for the
users/readers. The focus of our research is on single document summarisation. In
order to process a text document, it should be broken down into parts and then
represented in a suitable form to facilitate analysis. Various text representation schemes
have been studied, including n-gram, bag of words, and graphs. In our research we
use graphs to represent a text document. The graph is constructed by utilising
semantic relations such as dependency relations between words within the sentence.
2 Dense semantic graphs and its application in single document summarisation</p>
      <p>
        We propose a novel graph generation approach, which is an extension of an
existing semantic graph generation approach [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] by including more dependencies from
dependency parsing of the sentence. This results in dense semantic graph. We
evaluated both graphs in a text summarisation task through experiments. Results show that
our dense semantic graph outperformed the original semantic graph for unsupervised
extractive text summarization.
      </p>
      <p>The next section gives a short literature review of the earlier graph based
approaches to text summarisation. In section 3, a detailed description is provided
concerning the construction of two different semantic graphs that were used in our study.
Section 4 discusses extractive summarisation based on these semantic graphs and
section 5 describes the experiments and results. After that conclusion of the analysis
follows.
2</p>
    </sec>
    <sec id="sec-2">
      <title>Previous Work on Graph based Text Summarisation</title>
      <p>
        Earlier researchers have used graph representation of documents and properties of
graphs to extract important sentences from documents to create a short summary.
Graph based text summarisation methods such as LexRank [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], TextRank [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and
Opinosis [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] have shown good performance. There are two types of graph that are
constructed and used to represent text. Lexical graph uses the lexical properties of text
to construct a graph. LexRank and Text Rank are lexical graph based approaches.
They construct graphs by connecting two sentences/smaller text units as nodes in the
graph based on the degree of content overlap between them.
      </p>
      <p>On the other hand, semantic graph is based on semantic properties of text.
Semantic properties are: Ontological relationship between two words such as synonymy,
hyponymy; relationship among set of words representing the syntactic structure of
sentence such as dependency tree and syntactic trees. A set of words along with the
way they are arranged provides meaning. The same set of words connected in
different ways gives different meaning.</p>
      <p>
        According to the semantic properties utilised for graph construction, various
representations have been reported in literature for semantic graphs [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ]. Some of the
approaches utilize the lexical database WordNet to generate ontological relations
based semantic graph. In this sentences are broken into terms, mapped to WordNet
synsets and connected over WordNet relations [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]. In one of the approaches called
semantic Rank [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], sentences are connected as nodes and the weight of the edges
between them is the similarity score calculated by WordNet and Wikipedia based
similarity measures. Other approaches to generate semantic graphs try to utilize the
dependency relations of words in a sentence along with the ontological relations
between words. Utilizing this particular order of connection also forms the basis of
research work done on semantic graphs in our study. In this area of semantic graph
generation most of the work has been concentrated on identifying logical triples
(subject-object-predicate) from a document and then connecting these triples based on
various semantic similarity measures [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. Predicate (or verb) is the central part of any
sentence, which signifies the main event happening within the sentence. Thus it was
      </p>
      <p>Dense semantic graphs and its application in single document summarisation 3
mostly agreed to consider the verb and its main arguments (subject and object) as the
main information presented in the sentence, and use this as a basic semantic unit of
the semantic graph. Various researches have been done on this graph in the field of
supervised text summarisation.</p>
      <p>
        We have evaluated two semantic graphs which are based on the dependency
structure of words in a sentence. The first graph is triple(subject-object-verb) based
semantic graph proposed by Leskovec et al [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. The second graph is a novel approach of
semantic graph generation proposed in this paper, based on the dependency path
length between nodes. Our hypothesis is that moving to a dense semantic graph, as we
have defined it, is worthwhile. The principle idea behind this new graph has been
used in earlier research in kernel based relation identification [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. However it has not
been used for construction of a semantic graph for the complete document. The next
section describes more details about this graph.
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>Semantic Graphs</title>
      <p>In the research carried out in this paper, we have analysed the difference between
performances when more dependency relations than just subject-object-verb are
considered to construct a semantic graph of the document. In this direction, we have
developed a methodology to select the dependencies and nodes within a shortest
distance path of dependency tree to construct the semantic graph. First we will describe
the previous use of graphs and then we will introduce the graph generated by our
methodology.
3.1</p>
      <sec id="sec-3-1">
        <title>Semantic graph derived from a triplet (Subject-Object-verb)</title>
        <p>
          Leskovec et al. [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] has described this graph generation approach for their supervised
text summarization, where they train a classifier to learn the important relations
between the semantic graph of a summary and the semantic graph of an original text
document. In this graph the basic text unit is a triple extracted from sentence:
subjectverb-object. This is called triple as there are three connected nodes. Information such
as adjectives of subject/object nodes and prepositional information (time, location) are
kept as extra information within the nodes. After extracting triples from every
sentence of the text document two further steps are taken: i. co-reference and anaphora
resolution: all references to named entities (Person, Location etc.) and pronoun
refeences are resolved. ii. Triples are connected if their subject or object nodes are
synoymous or referring to the same named entity. Thus a connected semantic graph is
generated.
3.2
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>Dense Semantic graphs generated from shortest dependency paths between Nouns/Adjectives</title>
        <p>We have observed that various named entities such as location/time which are
important information, are not covered in the subject-predicate-object relations. As this
4 Dense semantic graphs and its application in single document summarisation
information is often added through prepositional dependency relations, it gets added
to nodes as extra information in the semantic graph generated by previous approaches.
However these named entities hold significant information to influence ranking of the
sentences for summary generation and to connect nodes in the semantic graph. This
has formed the basis of our research into new way of semantic graph generation. First
we elaborate the gaps observed in previous approach of semantic graph generation
and then give the details of the new semantic graph.</p>
      </sec>
      <sec id="sec-3-3">
        <title>Gaps identified in triple (subject-object-verb) based semantic graph.</title>
        <p>The kind of information loss observed in the previous semantic graphs has been
described below:
 Loss of links between words in sentence</p>
        <p>Some connections between named entities are not considered because they do not
come into the subject/object category. This information is associated with
subject/object, but does not get connected in the semantic graph, as they are not
directly linked through a predicate. For example consider the sentence below:
President Obama’s arrival in London created a joyful atmosphere.</p>
        <p>The triple extracted from this sentence is:</p>
        <p>Arrival-&gt;create-&gt;atmosphere
Here the information London, Obama is added as extra information to node
Arrival, and Joyful is added to node Atmosphere. However a direct link between
London and atmosphere is missing, whereas a reader can clearly see this is atmosphere
of London. This connection can be identified in our shortest dependency path
graph as shown below:</p>
        <p>London-prep-in-&gt;Arrival-nsubj-&gt;created-dobj-&gt;atmosphere
 Loss of inter-sentence links between words</p>
        <p>Some named entities which are not subject/object in one sentence are
subject/object of another sentence. When creating a semantic graph of complete
document, these entities are the connecting words between these sentences. In the
previous graph these connections are lost as shown below by two sentences.</p>
        <p>He went to church in Long valley.</p>
        <p>One of the explosions happened in Long Valley.</p>
        <p>The triple extracted from these sentences is:
He-&gt;went&gt;church
Explosion-&gt;happened-&gt;long valley</p>
        <p>Dense semantic graphs and its application in single document summarisation 5
In the semantic graph derived from triples of the above 2 sentences, we do not have
both these sentences connected, because the common link Long Valley is hidden as
extra information in one semantic graph.
 Identification of subject is not clear</p>
        <p>In a few cases, identification of a subject for the predicate is not very accurate with
current dependency parsers. This case occurs in the clausal complement of verb
phrase or adjectival phrases called dependency relation “xcomp”. Here the
determination of subject for clausal complement is not very accurate, as the subject is
external.</p>
      </sec>
      <sec id="sec-3-4">
        <title>Construction of shortest distance dependency path based semantic graph</title>
        <p>To overcome these gaps, we construct the semantic graph by connecting all noun and
adjectives which are connected within a shortest path distance in the dependency tree
of that sentence. From the literature review it has been identified that nouns are the
most important entities to be considered for ranking sentences. So we have decided to
include nouns as nodes in the semantic graph. We also considered adjectives, as they
modify nouns and may present significant information. The length of the shortest path
is varied from 2-5 to analyse its effect on the efficiency of the PageRank score
calculation. The following steps are followed to construct the semantic graph
 Co-reference resolution of named entities</p>
        <p>The text document is preprocessed to resolve all co-references of named entities.
We replace the references with the main named Entity for Person, Location, and
organization.
 Pronominal resolution</p>
        <p>After co-reference resolution, text is preprocessed for pronominal resolution. All
reference (he, she, it, who) are resolved to referring named entities and replaced
them in text.
 Identifying nodes and edges of the semantic graph</p>
        <p>The shortest path distance based Semantic graph is defined as G= (V, E), Where
{⋃
(
)
{
}}
In (1) ( ) provides part of the speech tag of . According to Penn tag
set for part of speech tags, “JJ” signifies Adjectives and “NN” signifies Noun.</p>
        <p>
          Edge set
{⋃
(
)
(
)
}
(1)
(2)
In (2) ( ) is the shortest distance from u to v in the dependency tree of that
sentence and limit is the maximum allowed shortest path distance, which is varied
from 2-5 in our experiments.
6 Dense semantic graphs and its application in single document summarisation
We have used Stanford CoreNLP package for co-reference resolution, identification
of named entities and dependency parse tree generation[
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. To develop the
graphs and calculate the page rank scores of nodes we use the JUNG software
package1. First we extract dependency relations for each sentence. Then we generate a
temporary graph for the dependency tree of that sentence in JUNG .Then Dijkstra’s
shortest path algorithm is applied to find the shortest distance between nodes. From
this temporary graph we find vertices and edges based on equations (1) and (2) to
construct the semantic graph.
        </p>
        <p>Fig. 1 and 2 show two graphs, triple based semantic graph and shortest distance
dependency path based semantic graphs for the given excerpt of 2 sentences below,
taken from the Long Valley document of DUC2002 data.</p>
        <sec id="sec-3-4-1">
          <title>A text excerpt taken from DUC 2002 data.</title>
          <p>The resort town's 4,700 permanent residents live in Long Valley, a 19-mile-long, 9-mile-wide
volcanic crater known as a caldera. Eruptions somewhat smaller than Mount St. Helens'
happened 550 years ago at the Inyo craters, which span Long Valley's north rim, and 650 years ago
at the Mono craters, several miles north of the caldera.
The next section describes the methodology to rank sentences based on the semantic
graph described in this section.
In this paper we want to analyse the impact of dense semantic graphs on text
summarisation and provide a comparison with the summarisation results of earlier triple
based semantic graphs. To achieve this, first we rank the semantic graph by one of the
graph ranking algorithm. We have used PageRank method to rank the semantic graph
nodes.</p>
          <p>The PageRank score of nodei is calculated as:
PageRank(nodei )  (1  d )  d *

node jIn(nodei )</p>
          <p>PageRank(node j )
|Out(node j )|
(3)
8 Dense semantic graphs and its application in single document summarisation</p>
          <p>Where d is the probability of jumping from nodei to any random node in the graph,
typically set between 0.1-0.2. In(nodei) is the set of incoming edges to nodei and
Out(nodej) is the set of outgoing edges of nodej. Initially PageRank of all nodes is
intialised with arbitrary values, as it does not affect the final values after convergence.
In this paper semantic graphs are undirected graphs so incoming edges of a node are
equal to outgoing edges.</p>
          <p>After calculating PageRank score of the nodes in the semantic graph, the score of
sentence Si in the text document is calculated by following equation:
∑
(
)
(4)
where nodej is the stemmed word/phrase in the graph representation. Scores are
normalised after dividing by the maximum score of sentences. After calculating
normalized scores of all sentences in the text document, sentences are ordered according to
their scores. As per the summary length, higher scoring sentences are taken as
summary sentences.</p>
          <p>In addition to this summary generation method, we have also tried to analyze impact
of including additional features together with PageRank scores on semantic graph
based text summarisation. This was done in a separate experimental run where we
have included sentence position as an additional feature for scoring of sentences.
Since the data we have experimented with is news data, a higher score is given to
early sentences of the document. So the score of a sentence Si after including sentence
position, i as a feature is given by:
newScore</p>
          <p>Si
 0.1 (Countsentences  i)</p>
        </sec>
        <sec id="sec-3-4-2">
          <title>Count sentences</title>
          <p> 0.9 ScoreSi
(5)</p>
          <p>After calculating the new score of the sentences, higher scoring sentences are
extracted as the summary as in previous summarisation method. The next section
describes the experimental setup.
5</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>Experiments</title>
      <p>We have experimented on two single document summarisation corpuses from
Document Understanding Conference (DUC), DUC-01 and DUC-02.</p>
      <p>
        DUC-01 contains 308 text documents and DUC-02 contains 567 text documents.
Both sets have 2 human written summaries per document for evaluation purposes. We
have used the ROUGE toolkit to evaluate system generated summaries with reference
summaries, that are 2 human generated summaries per document [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]. The ROUGE
toolkit has been used for DUC evaluations since the year 2004. It is a recall oriented
evaluation metric which matches n-grams between a system generated summary and
reference summaries.
      </p>
      <p>∑ {
∑ {
}∑
} ∑
(
(
)
)
(6)</p>
      <p>Dense semantic graphs and its application in single document summarisation 9
Rouge-1 is 1-gram metric. Rouge-2 is 2-gram metric. Rouge-W is the longest
weighted sequence metric, which gives weight to consecutive longest sequence
matches.</p>
      <p>
        ROUGE scores were calculated for different summarisation runs on triple based
semantic graphs and shortest dependency distance path based semantic graphs. On
triple based graphs two summarisation tasks were run for DUC01 and DUC-02 data.
The first considered PageRank only and the second used PageRank, sentence position
(Triple based, Triple + position). On the Shortest distance dependency path based
semantic graph, 6 summarisation tasks were run for both datasets. The first 4 runs are
based on PageRank scores alone by varying shortest distance from 2-5: shortest
distance 2 (SD-2), shortest distance 3 (SD-3), shortest distance 4 (SD-4) and shortest
distance 5 (SD-5). The fifth and sixth run include sentence position as feature with
SD-4 and SD-5(SD-4 + position, SD-5+ position). We have also compared our results
with the results of the text summarisation software Open Text Summarizer(OTS) [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ],
which is freely available and has been reported to perform best between other
available open source summarizers.
6
      </p>
    </sec>
    <sec id="sec-5">
      <title>Results and Analysis</title>
      <p>10 Dense semantic graphs and its application in single document summarisation
0.4
0.35</p>
      <p>0.3
0.25</p>
      <p>0.2
0.15</p>
      <p>0.1
0.05
0
ROUGE-1
ROUGE-2
ROUGE-w</p>
      <p>Including sentence position as a feature, improves the summarisation results on
both triple based graph and shortest distance dependency path based semantic graph.
Also in this case, ROUGE scores for summarisation run on shortest distance
dependency path based semantic graph are higher than for triple based semantic graphs. This
also indicates that we can include more features to improve the results further. Overall
results indicate that shortest distance based semantic graphs performs better in
ranking the sentences and are comparable to benchmark system OTS.</p>
      <p>Dense semantic graphs and its application in single document summarisation 11</p>
      <p>PageRank based summarisation is a novel approach for both our approaches.
Earlier for triple based semantic graph, PageRank node score was considered as a feature
for supervised text summarisation. In this paper we have looked at unsupervised
single document summarisation. In the evaluation, we have seen that only PageRank
based summarisation results do not exceed the benchmark results, but are comparable.
Benchmark OTS system utilises a language specific lexicon for identifying
synonymous words and cue terms. In future work, we can include a similar lexicon to
identify more relation between words to improve the performance. In this paper we have
hypothesised that if more dependency relations are considered for semantic graph
generation it gives better PageRank scores and thus improves the ranking accuracy for
extraction of summary sentences. Although triple based graphs are more visually
understandable they can be enhanced by adding more dependencies. When sentence
position was included as an extra feature, it improved the Rouge scores. Also it is
noticeable that summarisation results for shortest distance dependency path based
semantic graph are similar to results after including the additional feature sentence
position. This makes this graph equally useful in domains where sentence position
does not have an effect on importance.</p>
      <p>In future work we will apply semantic similarity and word sense disambiguation to
improve the connectivity of the graph and identify more relations between nodes.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <given-names>G.</given-names>
            <surname>Erkan</surname>
          </string-name>
          and
          <string-name>
            <given-names>D. R.</given-names>
            <surname>Radev</surname>
          </string-name>
          , “
          <article-title>LexRank: graph-based lexical centrality as salience in text summarization</article-title>
          ,
          <source>” Journal of Artificial Intelligence Research</source>
          , vol.
          <volume>22</volume>
          , no.
          <issue>1</issue>
          , pp.
          <fpage>457</fpage>
          -
          <lpage>479</lpage>
          , Jul.
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <given-names>R.</given-names>
            <surname>Mihalcea</surname>
          </string-name>
          and
          <string-name>
            <given-names>P.</given-names>
            <surname>Tarau</surname>
          </string-name>
          , “TextRank: Bringing Order into Texts,”
          <source>in Proceedings of Empirical Methods in Natural Language Processing</source>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <given-names>K.</given-names>
            <surname>Ganesan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zhai</surname>
          </string-name>
          , and J. Han, “
          <article-title>Opinosis : A Graph-Based Approach to Abstractive Summarization of Highly Redundant Opinions</article-title>
          ,”
          <source>in Proceedings of the 23rd International Conference on Computational Linguistics</source>
          ,
          <year>2010</year>
          , no.
          <source>August</source>
          , pp.
          <fpage>340</fpage>
          -
          <lpage>348</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <given-names>J.</given-names>
            <surname>Leskovec</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Milic-Frayling</surname>
          </string-name>
          , and
          <string-name>
            <given-names>M.</given-names>
            <surname>Grobelnik</surname>
          </string-name>
          , “
          <source>Extracting Summary Sentences Based on the Document Semantic Graph,” Microsoft Technical Report TR-2005-07</source>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <given-names>D.</given-names>
            <surname>Rusu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Fortuna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Grobelnik</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Mladenić</surname>
          </string-name>
          , “
          <article-title>Semantic Graphs Derived From Triplets With Application in Document Summarization,”</article-title>
          <source>Informatica Journal</source>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <given-names>L.</given-names>
            <surname>Plaza</surname>
          </string-name>
          and
          <string-name>
            <given-names>A.</given-names>
            <surname>Díaz</surname>
          </string-name>
          , “
          <article-title>Using Semantic Graphs and Word Sense Disambiguation Techniques to Improve Text Summarization,” Procesamiento del Lenguaje Natural Revista</article-title>
          , vol.
          <volume>47</volume>
          , pp.
          <fpage>97</fpage>
          -
          <lpage>105</lpage>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <given-names>G.</given-names>
            <surname>Tsatsaronis</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Varlamis</surname>
          </string-name>
          , and
          <string-name>
            <given-names>K.</given-names>
            <surname>Nørvåg</surname>
          </string-name>
          , “
          <article-title>SemanticRank: ranking keywords and sentences using semantic graphs</article-title>
          ,
          <source>” COLING '10 Proceedings of the 23rd International Conference on Computational Linguistics</source>
          , no.
          <source>August</source>
          , pp.
          <fpage>1074</fpage>
          -
          <lpage>1082</lpage>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <given-names>R. C.</given-names>
            <surname>Bunescu</surname>
          </string-name>
          and
          <string-name>
            <given-names>R. J.</given-names>
            <surname>Mooney</surname>
          </string-name>
          , “
          <article-title>A Shortest Path Dependency Kernel for Relation Extraction</article-title>
          ,”
          <source>in Proceedings of the conference on Human Language Technology and Empirical Methods in Natural Language Processing</source>
          ,
          <year>2005</year>
          , no.
          <source>October</source>
          , pp.
          <fpage>724</fpage>
          -
          <lpage>731</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Klein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. D.</given-names>
            <surname>Manning</surname>
          </string-name>
          , and
          <string-name>
            <given-names>Y.</given-names>
            <surname>Singer</surname>
          </string-name>
          , “
          <article-title>Feature-rich part-of-speech tagging with a cyclic dependency network,” in Proceedings of the 2003 Conference of the North American Chapter of the Association for Computational Linguistics on Human Language Technology -</article-title>
          <source>NAACL '03</source>
          ,
          <year>2003</year>
          , vol.
          <volume>1</volume>
          , pp.
          <fpage>173</fpage>
          -
          <lpage>180</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <given-names>H.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Peirsman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Chambers</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Surdeanu</surname>
          </string-name>
          , and
          <string-name>
            <given-names>D.</given-names>
            <surname>Jurafsky</surname>
          </string-name>
          , “
          <article-title>Stanford's multi-pass sieve coreference resolution system at the CoNLL-2011 shared task</article-title>
          ,”, Jun.
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <given-names>C.</given-names>
            <surname>Lin</surname>
          </string-name>
          and
          <string-name>
            <given-names>M.</given-names>
            <surname>Rey</surname>
          </string-name>
          , “R OUGE 
          <article-title>: A Package for Automatic Evaluation of Summaries,”</article-title>
          <source>in Proceedings of the ACL-04 Workshop: Text Summarization Branches Out</source>
          ,
          <year>2004</year>
          , pp.
          <fpage>74</fpage>
          -
          <lpage>81</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12. V. a. Yatsko and
          <string-name>
            <given-names>T. N.</given-names>
            <surname>Vishnyakov</surname>
          </string-name>
          , “
          <article-title>A method for evaluating modern systems of automatic text summarization</article-title>
          ,
          <source>” Automatic Documentation and Mathematical Linguistics</source>
          , vol.
          <volume>41</volume>
          , no.
          <issue>3</issue>
          , pp.
          <fpage>93</fpage>
          -
          <lpage>103</lpage>
          , Jun.
          <year>2007</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>