<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>To text summarization by dynamic graph mining</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Matej Gallo</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Luboš Popelínský</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Karel Vaculík</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>KD Lab FI MU Brno popel@fi.muni.cz</string-name>
        </contrib>
      </contrib-group>
      <pub-date>
        <year>2018</year>
      </pub-date>
      <volume>2203</volume>
      <fpage>28</fpage>
      <lpage>34</lpage>
      <abstract>
        <p>We show that frequent patterns can contribute to the quality of text summarization. Here we focus on single-document extractive summarization in English. Performance of the frequent patterns based model obtained with DGRMiner yields the most relevant sentences of all compared methods. Two out of three proposed methods outperform other methods if compared on ROUGE data.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1 Introduction</title>
      <p>
        Extractive summarization assigns a score to each text
unit (phrase, sentence, paragraph, passage) and then picks
the most informative ones to form a summary called
extract. The selected sentences are used verbatim [
        <xref ref-type="bibr" rid="ref1 ref25">1, 25</xref>
        ].
      </p>
      <p>Graph representation is a common way for
representing data in automatic text summarization tasks. We
introduce a new method for single-document extractive
summarization in English language where a text is represented
as a dynamic graph and each sentence corresponds to a
dynamic graph snapshot, i.e. the graph in a partivular time.
E.g. the first sentence is the oldest snapshot while the last
sentence of a text is the newest one. This method is based
on the principle of mining frequent patterns from such a
dynamic graph and using the resulting patterns as
indicators of sentence importance.</p>
      <p>The structure of this text is following. In Section 2.
we introduce DGRMiner, a tool for mining in dynamic
graphs. Section 3. describes the algorithm. In Sections
4 we briefly introduce the data used for evaluation.
Sections 5 and 6 explain a way of summarization evaluation
and sentence scoring. Section 7 displays the main results
of text summarization by means of frequent patterns mined
from dynamic graphs. Related work is a contents of
Section 8. We conclude with concluding Section 9.
ensure that only significant rules are extracted, the
algorithm incorporates measuring support and confidence.
While support expresses what portion of the graph is
affected by the rule, confidence measures the occurrence
frequency of a specific change, given that a particular
pattern was observed. Time abstraction is an extension to the
DGRMiner that allows us to analyse broader class of
dynamic graphs. The abstraction lies in the use of signum
function on relative timestamps. All the negative
timestamps become −1, all the positive timestamps become 1
and the timestamp of 0 remains 0. DGRMiner allows for
two types of abstraction. One affects only timestamps of
vertices and should be used when most changes are caused
by edges and vertices remain static. The second one also
affects the edges and is useful when there are too few
patterns with exact timestamps.
3</p>
    </sec>
    <sec id="sec-2">
      <title>Method</title>
      <p>The initial idea was to take a text as a temporal (i.e.
dynamic) graph where each sentence represent a graph
snapshot at a particular time and tokens (a lemma together with
a part-of-speech tag) were nodes and edges connected all
pairs of tokens. The label of an edge was equal to a
number of appearances of this pair of tokens. In the following
subsections we describe particular steps of the algorithm.
3.1</p>
      <sec id="sec-2-1">
        <title>Transforming Text to graphs</title>
        <p>
          For pre-processing we employed CoreNLP [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]. CoreNLP
provides a set of natural language analysis tools: POS
tagger, named entity recognition, parser, coreference
resolution system, sentiment analysis, bootstrapped pattern
learning, and open information extraction. We used four
annotators: sentence split, tokenize, lemma, and POS.
        </p>
        <p>Using the coreNLP package for R, we split the text into
sentences. In every iteration we removed the stop words
from a sentence, lemmatize the remaining words and
assigned the POS tags. If a lemma-tag pair was not already
in the graph, we added it in and created edges between all
words in a same sentence. Otherwise, we only notified the
DGRMiner that a particular word had appeared again. We
parametrized context c. Therefore, in each step we
modelled at most c sentences.</p>
        <p>We will show it on simple example that contains a single
sentence.
&lt;sentence position = " 9 " labelers
= "l, 2 , 3, 4"&gt;The great thing about
Aspen is that it has at least one of
each of the really useful stores that
you need, sellingstuff at regular
prices. &lt;/sentence&gt;
After removing stop words and labeling words with a
lemma-tag pair, we receive a graph t # 9.
t # 9
an 74 great#ADJ
an 75 thing#NOUN
an 11 Aspen#NOUN
an 76 least#ADJ
an 77 one#NUM
an 78 really#ADV
an 79 useful#ADJ
an 66 store#NOUN
an 5 need#VERB
ae u 48 5 11 [need#VERB]-[Aspen#NOUN]
ae u 432 5 66 [need#VERB]-[store#NOUN]
ae u 433 5 67 [need#VERB]-[sell#VERB]
ae u 434 5 74 [need#VERB]-[great#ADJ]
ae u 435 5 75 [need#VERB]-[thing#NOUN]
ae u 436 5 76 [need#VERB]-[least#ADJ]
ae u 437 5 77 [need#VERB]-[one#NUM]
In the first column, an, ae stand for "add a node" and
"add an edge" respectively.
3.2</p>
      </sec>
      <sec id="sec-2-2">
        <title>Pattern Mining</title>
        <p>Then we applied the DGRMiner on the data from
previous step to obtain predictive patterns, i.e. rules that
describe frequent (or rare) changes between past and future
graphs – sentences. We received two types of patterns:
frequent single-vertex patterns corresponding to nodes and
rare patterns corresponding to edges. We modified the
support parameter to change the threshold on frequency of
observed patterns. When the support parameter was set too
low, the DGRMiner considered a pattern anything that
appeared at least once in the text – every word, every possible
word combination. By setting the confidence parameter
in DGRMiner to 0, the DGRMiner assigned confidence
(as discussed in section 2) to each extracted pattern. We
also played with time abstraction which allowed us to
ignore the preset context c as discussed in section 2.
Therefore, patterns were observed in sentences that were more
than c − 2 sentences apart. Although this gave us more
patterns, many of them were uninformative. An example
of observed single-vertex pattern is "+supplies#NOUN".
This pattern indicates that the noun supplies appeared at
least n times in the text, where the n is determined by the
support parameter. An example of multi-vertex pattern is
[store#NOUN]-[sell#VERB], which tells us that the noun
store is frequently closely accompanied by the verb sell.
The proximity of these two words is given by context c
and the frequency is given by the support parameter.
4</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>Data</title>
      <p>
        To assess the performance of our method we used the Blog
summarization dataset [
        <xref ref-type="bibr" rid="ref10 ref11 ref23">10, 11, 23</xref>
        ]. It consists of 100 posts
annotated in XML that were randomly chosen from two
blogs (half from each blog), Cosmic Variance and
Internet Explorer Blog. Each of the four human
summarizers picked approximately 7 sentences to form 4 reference
summaries in total. We manually restored apostrophes for
shortened forms of to be and to have verbs, and in
possessive nouns. Punctuation within sentences was omitted as
the coreNLP sentence split annotator often wrongly split
the sentences in the middle. We decided not to use CNN
Dataset, nor SUMMAC dataset mentioned in [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] because
the provided reference summaries were not extracted,
verbatim sentences. This would require us to manually find
the best matching sentence from within the document.
5
      </p>
    </sec>
    <sec id="sec-4">
      <title>Semi-automatic evaluation</title>
      <p>
        Evaluating summaries on sentence level can be done
semiautomatically by measuring content overlap with
precision, recall, and F1 measure. An extracted sentence is
considered acceptable if the same sentence was extracted
in a reference summary. This process cannot be fully
automatized because reference summaries are created by
human judges. Other semi-automatic evaluation methods
used nowadays are: ROUGE, PYRAMID and BASIC
ELEMENTS. We will discuss only the ROUGE method [
        <xref ref-type="bibr" rid="ref14 ref25">25, 14</xref>
        ]
as it was used in this work.
5.1
      </p>
      <sec id="sec-4-1">
        <title>ROUGE</title>
        <p>The ROUGE (Recall-Oriented Understudy for Gisting
Evaluation) measure is based on the BLEU metrics used
in machine translation tasks. The idea is to compare the
differences between the distribution of words in the
candidate summary and the distribution of words in the
reference summaries. Given h reference summaries and a
candidate summary they are split into n-grams to calculate
the intersection of n-grams between the references and the
candidate. This process is illustrated in figure 1.</p>
        <p>
          Given its correlation with manual judgments ROUGE
almost seems to have become a standard. [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ] reports
Pearson coefficient for the most commonly used variations
(ROUGE-2 and ROUGE-SU4 [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]) at a value of between
0.94 and 0.99. Generally, the ROUGE-n is calculated from
the co-occurrences of n-grams between the candidate and
reference summaries as shown by formula 1 in [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ]:
ROUGE-n =
∑n-grams ∈ {Sumcan ∩ Sumref}
∑n-grams ∈ Sumref
(1)
where the numerator is the maximum number of
cooccurrences of n-grams in both reference and candidate
summary and the denominator is the total sum of the
number of n-grams present in the reference summaries.
        </p>
        <p>The ROUGE-SUγ is an adaptation of ROUGE-2 using
skip units (SU) of a size ≤ γ. SU4 considers the bigrams
and the bigrams SU to be arbitrary in a maximum window
length of γ = 4 words. The number of bigrams with an
arbitrary size of length γ is given by formula 2:
Count(k, n) = C
n
k</p>
        <p>k−γ
− ∑ (k − γ); γ ≤ 1
0
(2)
where n is the n-gram length and k is the sentence length
in words.</p>
        <p>
          The ROUGE metrics are not flawless. Firstly, they have
a problem with the representation of content. Secondly,
they will not consider chains of words such as "MUNI"
6= "Masaryk University" and "FI" 6=" Faculty of
Informatics". A study [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ] found that the system can be tricked into
generating a summary with high ROUGE score.
6
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Sentence Scoring</title>
      <p>In this step we used the observed patterns as indicators to
score the sentences. The final score was obtained
according to formula 3 as a sum of single-vertex and multi-vertex
scores:</p>
      <p>Score(s) = Scoresingle(s) + Scoremulti(s)
(3)
Three different metrics were implemented to calculate
single-vertex score.</p>
      <p>The Jaccard coefficient (JAC) as given by formula 4,
expresses the overlap of two sets. In our case, between the
set of patterns P and the set of words in the sentence S.</p>
      <p>ScoreJ(s) =
wsum(S ∩ P)
wsum(S ∪ P)
(4)
where wsum assigns weight to every element of the set and
then sums over them. A question arises regarding what
weight we should assign to non-indicators. Empirically,
we received the best results for the value of 0.001.</p>
      <p>The frequency method (FRQ) as given by formula 5,
is the simplest of the three. For a sentence s and a set of
patterns P the score is calculated as weighted average.</p>
      <p>ScoreF (s) = ∑ c(p)w(p)</p>
      <p>p∈P
where c(p) is the frequency of the pattern in the sentence
and w(p) is its associated weight.</p>
      <p>The density method (DEN) as given by formula 6,
counts the patterns in a sentence and normalizes by the
length of the sentence. This method is parameter-free.</p>
      <p>ScoreD(s) = | S ∩ P |
length(s)
(5)
(6)</p>
      <p>For multi-vertex patterns we tested frequency and
density methods. The only difference between the methods is
that instead of length of sentences we use the number of
all possible word pair combinations |S2| .</p>
      <p>We built the summary in a greedy fashion. In every
iteration we picked the highest scoring sentence. Patterns
observed in the sentence were penalized by a parameter
λ . The scores were recomputed and the process continued
until a desired number of sentences was picked or until all
the sentences were used. For every method we discovered
the optimal value of λ . We tuned the parameter on 10% of
the entire dataset. The optimal values for λ are presented
in the tables 2 and 1. The ordering of sentences in the
final summary maintains the relative ordering in the original
document.
7</p>
    </sec>
    <sec id="sec-6">
      <title>Results</title>
      <p>
        We evaluated the model using the ROUGE-1 metric,
which is recommended for short summaries [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. Every
blog post was compared against four reference summaries
as described in chapter 4. The results of all three methods
can be seen in tables 1, 2 and 3. The first three columns
denote whether time abstraction on both vertices and edges
is used, whether the patterns were used to score sentences,
and whether the initial weights were determined by the
confidence of DGRMiner for the given pattern,
respectively. The last three columns correspond to the ROUGE-1
metrics – precision (what portion of sentences we selected
were part of the reference summary), recall (what portion
of sentences in reference summary we extracted), and f1
measure (harmonic mean of precision and recall).
      </p>
      <p>The FRQ method marked the most accurate sentences
among all the presented algorithms as can be seen from
figure 2. JAC method picked fewer correct sentences than
FRQ method but still more than any traditional approach.
Both FRQ and DEN methods ranked first in terms of
precision.</p>
      <p>The frequency-based method incorporating patterns
with no confidence weighting (identified as FRQ) achieved
the highest recall, the highest f1 score, and the highest
binall</p>
      <p>F
F
F
T
F
T
T</p>
      <p>T
binall</p>
      <p>F
F
F
F
T
T
T</p>
      <p>T
binall</p>
      <p>F
F
T
T
patterns</p>
      <p>F
T
T
F
F
T
T</p>
      <p>F
patterns</p>
      <p>T
F
F
T
F
T
F</p>
      <p>T
patterns</p>
      <p>T
F
T
F
weights</p>
      <p>F
F
T
F
T
F
T</p>
      <p>T
weights</p>
      <p>
        F
T
F
T
F
F
T
T
number of correctly chosen sentences. The highest
precision was achieved by density-based method
incorporating patterns with no confidence weighting (identified as
DEN). We chose these two models and the highest scoring
Jaccard method for comparison together with algorithms
presented in article [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. We could see that FRQ behaves
similarly to Word frequency (WF) algorithm proposed in
the paper. This is not surprising, because the single
vertexfrequent patterns correspond to words that appear at least
n-times in the text. Where n is determined by the support
parameter in the DGRMiner tool as described in section
3.2. The multi-vertex patterns improved the performance
of the summarizer in density models and in one case (the
highest scoring model) in frequency model.
8
      </p>
    </sec>
    <sec id="sec-7">
      <title>Related work</title>
      <p>
        Kupiec et al. introduce in their work [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] method inspired
by Edmundson’s [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]. They approached the summarization
as a statistical classification problem. A Bayes classifier
was trained to estimate the probability that a given
sentence would be included in the summary. They used 6
discrete features (presented in order of importance):
paragraph feature (the position of sentence in paragraph s),
fixed-phrase feature (the sentence contains a phrase from
a list), sentence length cutoff feature (threshold u1 = 5,
length(s) &gt; u1), thematic word feature (the presence of
thematic terms), uppercase word feature (the presence of
words in capital letters). The best results were obtained
using the first three features.
      </p>
      <p>
        Aone et al. [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] built on Kupiec’s work [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] and
expanded the feature set of their system, called DIMSUM,
with signature terms, which indicate key concepts for a
given document. Another advantage over [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ]’s system
is the use of multi-word phrases – statistically derived
collocation phrases (e.g. "omnibus bill", "crime bill",
"brady bill") and associated words ("camera" and
"obscura", "Columbia River" and "gorge"), as well as the use
of WordNet [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] to identify possible synonyms of found
signature terms. He applied a shallow discourse analysis to
resolve co-references and maintain cohesion – only name
aliases were resolved such as UK to United Kingdom.
      </p>
      <p>
        Osborne in his work [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] disagrees with the traditional
assumption of feature independence and shows
empirically that the maximum entropy (MaxEnt) model produces
better extracts than the naïve Bayes model with similarly
optimized prior appended to both models. Unlike naïve
Bayes, MaxEnt does not make unnecessary feature
independence assumptions. Let c be a binary label (binary:
part of summary or not), s the item we are interested in
labeling, fi the i-th feature, and ωi the corresponding feature
weight.
      </p>
      <p>
        Hidden Markov Models (HMM), similar to MaxEnt,
have weaker assumptions of independence. There are
three types of dependencies: positional dependence,
feature dependence, and Markovity dependence. A first-order
Markov model allows modeling these dependencies.
Conroy and O’leary [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] use a joint distribution for the features
set, unlike the independence-of-features assumption used
in naïve Bayesian methods. The HMM was trained
using five features: position of the sentence in the document
(number of states); number of terms in a sentence, and
likeliness of the sentence terms given the document terms.
      </p>
      <p>
        Summarizer NetSum presented by Svore et al. [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ] uses
an artificial neural network (ANN) called RankNet to rank
the sentences. RankNet is a pair-based neural network
algorithm for ranking a set of inputs. It is trained on pairs
of sentences (Si, S j), such that the ROUGE score for Si
should be higher than S j. Pairs are only generated in a
single document, not across documents. The cost function
for RankNet is the probabilistic cross-entropy cost
function. Training is performed using a modified version of
back-propagation algorithm for two-layer networks, which
is based on optimizing the cost function by gradient
descent. The system significantly outperforms the standard
baseline in the ROUGE-1 measure. No past system could
outperform the baseline with statistical significance.
      </p>
      <p>
        A system for generating product category-based
(topicbased) extractive summarization was proposed by [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ].
The collection of 45 news items corresponding to various
products are pre-processed using standard techniques:
tokenization, stopword removal, stemming. The final corpus
contains around 1500 features and is represented by a
bagof-words VSM based on these features. To identify the
topics, the news items about specific categories of
products are segregated into separate clusters using K-Means
and then an extractive summary is generated from each of
these topical clusters. The K number of cluster is
determined by a Self-Organizing Map (SOM).
      </p>
      <p>
        Chakraborti and Dey [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] assigned a score to the
entire summary as a single unit. The total summary score
(TSS) is taken as a combination of cosine similarity
between centroid of corpus and the summary as a whole;
relative length of the summary; and redundancy penalty.
To maximize TSS constrained by the number of lines in
summary (τ = 5 − 7%), they opted for quick Artificial Bee
Colony optimization, a global optimization technique, for
sentence selection. The summary with the highest score is
then chosen.
      </p>
      <p>
        In Muresan’s paper [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ], a system called GIST-IT used
for email-summarization task is discussed. First, noun
phrases (NPs) are extracted as they carry the most
contentful information. Subsequently, machine learning is used to
select the most salient NPs. A set of nine features, divided
into three categories (head of the NP, whole NP,
combination of head and modifiers of NP) were used: head of the
NP TF*IDF, position of first occurrence (focc) of the head
in text, TF*IDF of entire NP, focc of entire NP, length of
the NP in words, length of the NP in characters, position
of the NP in the sentence, position of the NP in the
paragraph, and combination of the TF*IDF scores of head of
the NP and its modifiers.
      </p>
      <p>To find the most salient NPs, three machine learning
algorithms were applied: decision trees (axis-parallel trees –
C4.5 and oblique trees – OC1), rule induction (production
rule – C4.5rules and propositional rules – RIPPER), and
decision forests (DFC using information gain ratio).</p>
      <p>
        Muresan claims that shallow linguistic filtering applied
to NPs improved the classifiers accuracy [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. The
filtering consisted of four steps: grouping inflectional variants,
removing unimportant modifiers, filtering stopwords, and
removing empty nouns.
      </p>
      <p>
        A modification of PageRank algorithm called LexRank
is used to weight sentences. The undirected graph of
sentences is constructed from symmetrical similarities
(modified cosine measure). The score of each vertex s is
calculated iteratively until the values of the vertices have not
been modified by more than ε = 0.0001. LexRank
algorithm is used as a component of the MEAD [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ].
      </p>
      <p>Unlike LexRank, TextRank uses the similarities of
edges to weight the vertices. The score of each sentence si
is calculated iteratively until convergence is reached.</p>
      <p>
        As the graph is constructed from inter-sentence
similarity measure, the choice of the method for
sentenceweighting has significant impact. One approach is to use
a bag-of-words to represent the sentences. The similarity
is obtained by calculating the cosine similarity weighted
by inverse document frequency [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] between their vectorial
representations. Another approach suggests using word
overlap between sentences instead. The weak point of all
similarity measures that use words (cosine similarity, word
overlap, longest common subsequence) is the dependency
on the lexicon of the document. The solution to this is
its combining with similarity measure based on chains of
characters. Therefore, sentences that do not share a single
word but contain a number of words that are close
morphologically can be compared [
        <xref ref-type="bibr" rid="ref25">25</xref>
        ].
      </p>
      <p>
        Patil [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] proposed a new graph-based model called
SUMGRAPH. First the text is pre-processed (stemmed)
and represented as a VSM with TF*IDF weights. He then
computes pair-wise cosine similarities and subtracts the
values from 1 to obtain dissimilarities. The resulting
matrix of intra-sentence dissimilarities is then used to model
the document as graph. The vertices represent the
sentences and edges are weighted by intra-sentence
dissimilarities.
      </p>
      <p>
        The novel idea is the use of link reduction technique
known as Pathfinder Network Scaling (PFnet) [
        <xref ref-type="bibr" rid="ref20 ref21">21, 20</xref>
        ] to
scale the graph. PFnet models the human aspects of
semantic memory. The centrality of a sentence and its
position in a document are used to compute the importance
of sentences. Four different centrality measure were tested
and closeness centrality showed to perform best. Finally,
the sentences are ranked according to their importance and
firstn highest-scoring sentences were picked.
      </p>
      <p>
        The approaches based on similarity graphs solely model
the similarity between pairs of sentences with no clear
representation of word relations. Therefore, it is not clear
if they adequately cover all topical information. The
hypergraph-based approach is to remedy this problem by
capturing the high-order relations between both sentences
and words. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] proposed a hypergraph-based model for
generic summarization based on [
        <xref ref-type="bibr" rid="ref27">27</xref>
        ]’s hypergraph-based
model for query-focused summarization. The ranking
uses a semi-supervised approach to order sentences. They
model the words as vertices and sentences as hyperedges
and then approach the problem as a random walk over
hyperedges.
9
      </p>
    </sec>
    <sec id="sec-8">
      <title>Conclusion</title>
      <p>We showed that frequent patterns can contribute to the
quality of text summarization. The comparison supports
our statement that the performance of our frequent patterns
based model is comparable to the simpler word frequency
method and yielded the most relevant sentences of all
compared methods. Our methods outperformed other methods
in precision but lacked in recall. We attribute the similarity
to word frequency method to inadequate graph
representation – instead of interconnecting all the words within a
sentence, suggest connecting them according to the parse
tree. Another consideration is to use a graph mining tool
that searches for more specific types of patterns than
DGRMiner.</p>
    </sec>
    <sec id="sec-9">
      <title>Acknowledgments</title>
      <p>This work has been supported by Faculty of Informatics,
Masaryk University. We would like to thank to ITAT
reviewers for their suggestions and comments.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Charu</surname>
            <given-names>C.</given-names>
          </string-name>
          <article-title>AGGARWAL and ChengXiang ZHAI</article-title>
          .
          <source>Mining Text Data</source>
          . Springer, New York,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Chinatsu</surname>
            <given-names>AONE</given-names>
          </string-name>
          ,
          <string-name>
            <surname>James</surname>
            <given-names>GORLINSKY</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Bjornar LARSEN</surname>
          </string-name>
          .
          <article-title>A trainable summarizer with knowledge acquired from robust nlp techniques</article-title>
          .
          <source>Advances in Automatic Text Summarization</source>
          ,
          <volume>71</volume>
          ,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Abdelghani</surname>
            <given-names>BELLAACHIA</given-names>
          </string-name>
          and
          <article-title>Mohammed ALDHELAAN. Multi-document hyperedge-based ranking for text summarization</article-title>
          .
          <source>In Proceedings of the 23rd ACM International Conference on Conference on Information and Knowledge Management</source>
          , pages
          <fpage>1919</fpage>
          -
          <lpage>1922</lpage>
          , New York,
          <year>2014</year>
          [cit. 2016-
          <volume>05</volume>
          -10]. ACM.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Swapnajit</given-names>
            <surname>CHAKRABORTI</surname>
          </string-name>
          <article-title>. Multi-document text summarization for competitor intelligence: A methodology based on topic identification and artificial bee colony optimization</article-title>
          .
          <source>In Proceedings of the 30th Annual ACM Symposium on Applied Computing</source>
          , pages
          <fpage>1110</fpage>
          -
          <lpage>1111</lpage>
          , New York,
          <year>2015</year>
          [cit. 2016-
          <volume>05</volume>
          -10]. ACM.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Swapnajit</surname>
            <given-names>CHAKRABORTI</given-names>
          </string-name>
          and
          <string-name>
            <surname>Shubhamoy DEY</surname>
          </string-name>
          .
          <article-title>Product news summarization for competitor intelligence using topic identification and artificial bee colony optimization</article-title>
          .
          <source>In Proceedings of the 2015 Conference on Research in Adaptive and Convergent Systems</source>
          , pages
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          , New York,
          <year>2015</year>
          [cit. 2016-
          <volume>05</volume>
          -10]. ACM.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>John</surname>
            <given-names>M. CONROY</given-names>
          </string-name>
          and
          <string-name>
            <surname>Dianne P. O'LEARY.</surname>
          </string-name>
          <article-title>Text summarization via hidden markov models</article-title>
          .
          <source>In Proceedings of the 24th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          , pages
          <fpage>406</fpage>
          -
          <lpage>407</lpage>
          , New York,
          <year>2001</year>
          [cit. 2016-
          <volume>05</volume>
          -10]. ACM.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Harold</surname>
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>EDMUNDSON</surname>
          </string-name>
          .
          <article-title>New methods in automatic extracting</article-title>
          .
          <source>J. ACM [online]</source>
          ,
          <volume>16</volume>
          (
          <issue>2</issue>
          ):
          <fpage>264</fpage>
          -
          <lpage>285</lpage>
          , april
          <year>1969</year>
          [cit. 2016-
          <volume>05</volume>
          -10].
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Günes</surname>
            <given-names>ERKAN</given-names>
          </string-name>
          and
          <string-name>
            <surname>Dragomir</surname>
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>RADEV</surname>
          </string-name>
          . Lexrank:
          <article-title>Graph-based lexical centrality as salience in text summarization</article-title>
          .
          <source>J. Artif. Int. Res. [online]</source>
          ,
          <volume>22</volume>
          (
          <issue>1</issue>
          ):
          <fpage>457</fpage>
          -
          <lpage>479</lpage>
          , december
          <year>2004</year>
          [cit. 2016-
          <volume>05</volume>
          -11].
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Rafael</surname>
            <given-names>FERREIRA</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Luciano</surname>
            <given-names>CABRAL</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Rafael</surname>
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>LINS</surname>
          </string-name>
          .
          <article-title>Assessing sentence scoring techniques for extractive text summarization</article-title>
          .
          <source>Expert Systems with Applications [online]</source>
          ,
          <volume>40</volume>
          (
          <issue>14</issue>
          ):
          <fpage>5755</fpage>
          -
          <lpage>5764</lpage>
          ,
          <year>october 2013</year>
          [cit. 2016-
          <volume>05</volume>
          - 10].
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Meishan</surname>
            <given-names>HU</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aixin</surname>
            <given-names>SUN</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Ee-Peng LIM</surname>
          </string-name>
          .
          <article-title>Commentsoriented blog summarization by sentence extraction</article-title>
          .
          <source>In Proceedings of the Sixteenth ACM Conference on Conference on Information and Knowledge Management</source>
          , pages
          <fpage>901</fpage>
          -
          <lpage>904</lpage>
          , New York,
          <year>2007</year>
          [cit. 2016-
          <volume>05</volume>
          -10]. ACM.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Meishan</surname>
            <given-names>HU</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Aixin</surname>
            <given-names>SUN</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Ee-Peng LIM</surname>
          </string-name>
          .
          <article-title>Commentsoriented document summarization: Understanding documents with readers' feedback</article-title>
          . In
          <string-name>
            <surname>Sung-Hyon</surname>
            <given-names>MYAENG</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Douglas</surname>
            <given-names>W. OARD</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fabrizio</surname>
            <given-names>SEBASTIANI</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tat-Seng</surname>
            <given-names>CHUA</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Mun-Kew</surname>
            <given-names>LEONG</given-names>
          </string-name>
          , editors,
          <source>Proceedings of the 31st Annual International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          , pages
          <fpage>291</fpage>
          -
          <lpage>298</lpage>
          , New York,
          <year>2008</year>
          [cit. 2016-
          <volume>05</volume>
          -10]. ACM.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Julian</surname>
            <given-names>KUPIEC</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jan</surname>
            <given-names>PEDERSEN</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Francine CHEN</surname>
          </string-name>
          .
          <article-title>A trainable document summarizer</article-title>
          .
          <source>In Proceedings of the 18th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          , pages
          <fpage>68</fpage>
          -
          <lpage>73</lpage>
          , New York,
          <year>1995</year>
          [cit. 2016-
          <volume>05</volume>
          -10]. ACM.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Chin-Yew</surname>
            <given-names>LIN</given-names>
          </string-name>
          .
          <article-title>Rouge: a package for automatic evaluation of summaries</article-title>
          . In Stan SZPAKOWICZ MarieFrancine MOENS, editor,
          <source>Workshop Text summarization Branches Out (ACL '04) [online]</source>
          , pages
          <fpage>74</fpage>
          -
          <lpage>81</lpage>
          , Barcelona,
          <year>2004</year>
          [cit. 2016-
          <volume>05</volume>
          -10]. ACL.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Petr</surname>
            <given-names>MACHOVEC</given-names>
          </string-name>
          .
          <article-title>Automatická sumarizace textu [online]</article-title>
          .
          <source>Master's thesis</source>
          , Masaryk University, Faculty of Informatics, Brno,
          <year>2015</year>
          [cit. 2016-
          <volume>05</volume>
          -10].
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Christopher</surname>
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Manning</surname>
            , Mihai Surdeanu, John Bauer, Jenny Finkel,
            <given-names>Steven J.</given-names>
          </string-name>
          <string-name>
            <surname>Bethard</surname>
          </string-name>
          , and
          <string-name>
            <surname>David McClosky</surname>
          </string-name>
          .
          <article-title>The Stanford CoreNLP natural language processing toolkit. In Association for Computational Linguistics (ACL) System Demonstrations</article-title>
          , pages
          <fpage>55</fpage>
          -
          <lpage>60</lpage>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>George</surname>
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>MILLER</surname>
          </string-name>
          .
          <article-title>Wordnet: A lexical database for english</article-title>
          .
          <source>Commun. ACM [online]</source>
          ,
          <volume>38</volume>
          (
          <issue>11</issue>
          ):
          <fpage>39</fpage>
          -
          <lpage>41</lpage>
          , november
          <year>1995</year>
          [cit. 2016-
          <volume>05</volume>
          -12].
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Smaranda</surname>
            <given-names>MURESAN</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Evelyne</surname>
            <given-names>TZOUKERMANN</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Judith</surname>
            <given-names>L. KLAVANS.</given-names>
          </string-name>
          <article-title>Combining linguistic and machine learning techniques for email summarization</article-title>
          .
          <source>In Proceedings of the 2001 Workshop on Computational Natural Language Learning - Volume 7</source>
          , pages
          <fpage>19</fpage>
          :
          <fpage>1</fpage>
          -
          <lpage>19</lpage>
          :
          <fpage>8</fpage>
          ,
          <string-name>
            <surname>Stroudsburg</surname>
          </string-name>
          ,
          <year>2001</year>
          [cit. 2016-
          <volume>05</volume>
          -10].
          <article-title>Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Miles</surname>
            <given-names>OSBORNE.</given-names>
          </string-name>
          <article-title>Using maximum entropy for sentence extraction</article-title>
          .
          <source>In Proceedings of the ACL'02 Workshop on Automatic Summarization, Stroudsburg</source>
          ,
          <year>2002</year>
          [cit. 2016-
          <volume>05</volume>
          - 10].
          <article-title>Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Kaustubh</surname>
            <given-names>PATIL</given-names>
          </string-name>
          and
          <string-name>
            <surname>Pavel BRAZDIL</surname>
          </string-name>
          .
          <article-title>Text summarization: Using centrality in the pathfinder network</article-title>
          .
          <source>Int. J. Comput. Sci. Inform</source>
          .
          <source>Syst [online]</source>
          ,
          <volume>2</volume>
          :
          <fpage>18</fpage>
          -
          <lpage>32</lpage>
          ,
          <year>2007</year>
          [cit. 2016-
          <volume>05</volume>
          -12].
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <surname>Roger</surname>
            <given-names>W. SCHVANEVELDT</given-names>
          </string-name>
          , editor.
          <source>Pathfinder Associative Networks: Studies in Knowledge Organization. Ablex Publishing Corp., Norwood</source>
          ,
          <year>1990</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>Roger</surname>
            <given-names>W.</given-names>
          </string-name>
          <string-name>
            <surname>SCHVANEVELDT</surname>
            ,
            <given-names>D.W.</given-names>
          </string-name>
          <string-name>
            <surname>DEARHOLT</surname>
            , and
            <given-names>F.T.</given-names>
          </string-name>
          <string-name>
            <surname>DURSO</surname>
          </string-name>
          .
          <article-title>Graph theoretic foundations of pathfinder networks</article-title>
          .
          <source>Computers &amp; mathematics with applications [online]</source>
          ,
          <volume>15</volume>
          (
          <issue>4</issue>
          ):
          <fpage>337</fpage>
          -
          <lpage>345</lpage>
          ,
          <year>1988</year>
          [cit. 2016-
          <volume>05</volume>
          -11].
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <surname>Jonas</surname>
            <given-names>SJÖBERGH</given-names>
          </string-name>
          .
          <article-title>Older versions of the rougeeval summarization evaluation system were easier to fool</article-title>
          . Inf. Process. Manage.,
          <volume>43</volume>
          (
          <issue>6</issue>
          ):
          <fpage>1500</fpage>
          -
          <lpage>1505</lpage>
          , november
          <year>2007</year>
          [cit. 2016-
          <volume>05</volume>
          -12].
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <surname>Dr</surname>
          </string-name>
          .
          <source>Aixin SUN. Commentsoriented document summarization</source>
          ,
          <year>2015</year>
          . http://www.ntu.edu.sg/home/axsun/datasets.html, visited
          <year>2016</year>
          -
          <volume>05</volume>
          -20.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <surname>Krysta</surname>
            <given-names>M. SVORE</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lucy</surname>
            <given-names>VANDERWENDE</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Christopher</surname>
            <given-names>J.C.</given-names>
          </string-name>
          <string-name>
            <surname>BURGES</surname>
          </string-name>
          .
          <article-title>Enhancing single-document summarization by combining RankNet and third-party sources</article-title>
          .
          <source>In Proceedings of the 2007 Joint Conference on Empirical Methods in Natural Language Processing and Computational Natural Language Learning (EMNLP-CoNLL)</source>
          , pages
          <fpage>448</fpage>
          -
          <lpage>457</lpage>
          , Prague,
          <year>2007</year>
          [cit. 2016-
          <volume>05</volume>
          -10].
          <article-title>Association for Computational Linguistics</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <surname>Juan-Manuel</surname>
            <given-names>TORRES</given-names>
          </string-name>
          -MORENO.
          <article-title>Automatic Text Summarization</article-title>
          . ISTE, London,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>Karel</given-names>
            <surname>Vaculík</surname>
          </string-name>
          and
          <string-name>
            <given-names>Lubomír</given-names>
            <surname>Popelínský</surname>
          </string-name>
          . Dgrminer:
          <article-title>Anomaly detection and explanation in dynamic graphs</article-title>
          . In
          <string-name>
            <surname>Knobbe-A. Soares C. Papapetrou P. Boström</surname>
          </string-name>
          , H., editor,
          <source>Advances in Intelligent Data Analysis XV - 15th International Symposium, IDA 2016</source>
          , pages
          <fpage>308</fpage>
          -
          <lpage>319</lpage>
          , Neuveden,
          <year>2016</year>
          . Springer.
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <surname>Wei</surname>
            <given-names>WANG</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Furu</surname>
            <given-names>WEI</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wenjie</surname>
            <given-names>LI</given-names>
          </string-name>
          ,
          <article-title>and Sujian LI. Hypersum: Hypergraph based semi-supervised sentence ranking for query-oriented summarization</article-title>
          .
          <source>In Proceedings of the 18th ACM Conference on Information and Knowledge Management</source>
          , pages
          <fpage>1855</fpage>
          -
          <lpage>1858</lpage>
          , New York,
          <year>2009</year>
          [cit. 2016-
          <volume>05</volume>
          -10]. ACM.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>