<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Towards Efficient model for Automatic Text Summarization</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Yetunde O. Folajimi</string-name>
          <email>yetundeofolajimi@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Tijesuni I. Obereke</string-name>
          <email>tobereke@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, University of Ibadan.</institution>
          ,
          <addr-line>+2348056648530</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Computer Science, University of Ibadan.</institution>
          ,
          <addr-line>+2348137462256</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2016</year>
      </pub-date>
      <fpage>7</fpage>
      <lpage>9</lpage>
      <abstract>
        <p>ive method of summarization involves the use of linguistic methods to analyze and interpret a document, the system then looks for another way to portray the content of the document in a short form and still pass across the main gist of the document. Also the input of a text summarization system can either be single or multiple. Single document summarization involves summarizing a single text while Multi-document summarization involves summarizing from more than one source text. Automatic text summarization is one of the many applications of Natural Language Processing. It can be used for question and answering, information retrieval among other things. Earlier methods of text summarization used statistical methods that assigned scores to sentences or words in a sentence, and these methods are inefficient because they didn't consider the context of words, which made the resulting summaries, incoherent. More research unveiled approaches that do not score sentences for extraction, but merged lots of knowledge bases to enable them know the part of speech of words in a sentence but do not consider keywords identification to identify important parts of documents.. Automatic text summarization system helps saves time and effort that one would have used to scan a whole document, it also helps increase productivity and with the amount of research that has been done in automatic text summarization, summaries are available in different languages [1]. This paper presents the current technologies and techniques as well as prevailing challenges in automatic text summarization, consequently, we propose a model for improving text summarization by using a method that combines sentence scoring algorithm with sentence reduction.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Automatic text summarization</kwd>
        <kwd>extractive summary</kwd>
        <kwd>sentence scoring</kwd>
        <kwd>sentence reduction</kwd>
        <kwd>query-based summarization</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. INTRODUCTION</title>
      <p>The area of automatic text summarization has become a widely
explored area of research because of the need for immediate
access to information at this age where the amount of information
on the World Wide Web is voluminous. The problem is not the
availability of information but users have access to more than
enough information than they need, they are also faced with the
problem of digging through that large amount of information to
get what they really need.</p>
      <p>Automatic text summarization is a process whereby a computer</p>
    </sec>
    <sec id="sec-2">
      <title>2. SENTENCE EXTRACTION METHODS</title>
    </sec>
    <sec id="sec-3">
      <title>FOR TEXT SUMMARIZATION</title>
      <p>
        A system that scans a document in machine-readable form then
selects from the sentences in the Article the ones that carry
important information was proposed in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. The significance
factor of a sentence is derived from an analysis of its words,
whereby the frequency of words occurrence is a useful
measurement of word significance and that the relative position of
words within a sentence having given values of significance
furnishes a useful measurement for determining the significance
of sentences. In [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], a justification was made about measure of
significance based on how frequent some words occurred by
pointing out that an author when trying to express his thoughts on
a subject repeats some words.
      </p>
      <p>
        Another research in IBM pointed out that the position of a
sentence can be used to find areas of a document containing
important information [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. There, it was shown that sentences that
occur in the initial or final parts of a paragraph contain important
information. By analyzing a sample of 200 paragraphs, it was
discovered that in most paragraphs the headings came first and in
few it came last.
      </p>
      <p>
        Unlike the method used in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], which used only the frequency of
word occurrence to produce extracts, [4] analyzed using cue
words, title and heading words, sentence location and key method
individually and together. The justification of using cue method is
that sentences containing words like “most importantly”, “in this
paper” indicate sentence importance. For key method, scores were
assigned to frequently occurring words in the document. For title
method, sentences are scored based on how much of the title or
heading words it contains and for the sentence location, the
importance of a sentence is determined using position as criteria
like words at the beginning of a paragraph are considered
important. His results showed that the best match between
automatic and human-written abstracts was accomplished when
sentence location, cue words and title words are considered.
2.1
      </p>
    </sec>
    <sec id="sec-4">
      <title>Beyond Sentence Extraction</title>
      <p>
        A method that involved the removal of irrelevant phrases from
sentences extracted for summary was introduced in [
        <xref ref-type="bibr" rid="ref4">5</xref>
        ]. The first
step involves the generation of a parse tree, followed by grammar
checking so as to know which of the nodes of the tree can be
deleted, it then checks the parts of the sentences that contains
information relating to the main topic. After doing all the above it
then removes the unnecessary parts of the sentences leaving
behind a concise and coherent summary.
      </p>
      <p>
        Motivated by the fact that automatic summarizers cannot always
identify where the main gist of a document lies and the way text is
generated is poor, [
        <xref ref-type="bibr" rid="ref5">6</xref>
        ] introduced a cut and paste method which
involved six operations:
      </p>
    </sec>
    <sec id="sec-5">
      <title>MACHINE LEARNING METHODS</title>
      <p>Various machine learning techniques have been exploited in
automatic text summarization. Some of the techniques used
include: Naïve-Bayes method, Rich Features and Decision Trees
method, Hidden Markov model, Log-linear models and Neural
Networks and Third Party Features.
3.1</p>
    </sec>
    <sec id="sec-6">
      <title>Naïve-Bayes Method</title>
      <p>
        Naïve-Bayes method was first used in [
        <xref ref-type="bibr" rid="ref6">7</xref>
        ] by using Bayesian
classifier to determine if a sentence should be extracted or not.
The system was able to learn from data. Some features used by
their system include the presence of uppercase words, length of
sentence, structure of phrase and position of words. The author
assumed the following:
s = a certain sentence, S = the sentences in the summary, and F1, ,
Fk = the features.
-- (1)
In equation 1, Sentences are scored based on these features and
the formula is used to calculate the score, the highest ranking
sentences are extracted.
      </p>
      <p>
        Th naïve-bayes classifier was also used in DimSum [
        <xref ref-type="bibr" rid="ref7">8</xref>
        ], which
used term frequency (tf) which is the number of times that a word
appears in a sentences and inverse document frequency (idf)
which is the number of sentences in which a word occurs, to know
words that hold point at the key concepts of a document.
3.2
      </p>
    </sec>
    <sec id="sec-7">
      <title>Rich Features and Decision Trees</title>
      <p>Decision trees are powerful and popular tools for classification
and prediction. It is a classifier in the form of a tree structure. The
following nodes make up the tree:



</p>
      <p>Decision node: specifies a test on a single attribute,
Leaf node: indicates the value of the target attribute,
Arc/edge: split of one attribute,</p>
      <p>
        Path: a disjunction of test to make the final decision
In [
        <xref ref-type="bibr" rid="ref8">9</xref>
        ], the authors concentrated on text position by making an
effort to determine how sentence position affects the selection of
sentences. The justification for the focus on position method is
that texts are in a particular discourse structure, and that sentences
containing ideas related to the topic of a document are always in
specifiable locations (e.g. title, abstracts, etc). They also
mentioned that discourse structure significantly varies over
domains, so therefore the position method cannot be easily
defined.
      </p>
      <p>
        A sentence reduction algorithm that is based on decision tree was
introduced in [
        <xref ref-type="bibr" rid="ref9">10</xref>
        ]. The algorithm proposed used semantic
information to aid the process of sentence reduction and decision
tree to handle the fact that the orders of original sentences change
after they are reduced. They extended Knight and Marcu’s
sentence compression algorithm [
        <xref ref-type="bibr" rid="ref10">11</xref>
        ], which was also based on
decision tree by adding semantic information to theirs. To achieve
this, they used a Parser to parse the original sentences and by
using WordNet, they enhanced the syntax tree gotten with
semantic information.
3.3
      </p>
    </sec>
    <sec id="sec-8">
      <title>Hidden Markov Models</title>
      <p>
        A hidden Markov model is a tool for denoting probability
distributions over sequences of observations. If we represent the
observation at time t by the variable Yt, we assume that the
observation are sampled at discrete, equally-spaced time intervals,
so t can be an integer-valued time index. The two defining
properties of hidden Markov model are: the assumption that the
observation at time t was generated by some process whose state
St is hidden from the observer and the assumption that the state of
the hidden process satisfies the Markov property i.e. given the
value of St-1, the current state St is independent of all the states
prior to t-1 [
        <xref ref-type="bibr" rid="ref11">12</xref>
        ].
      </p>
      <p>
        Two sentence reduction algorithms were proposed in [
        <xref ref-type="bibr" rid="ref12">13</xref>
        ]. Both
were template-translation based which means that they don’t need
syntactic parser to represent the original sentences for reduction.
One was founded on example-based machine-translation which
does a good job of in the area of sentence reduction. On the other
hand in specific cases, the computational complexity can be
exponential. While the second one was an addition to the
template-translation algorithm through the application of Hidden
Markov model, the model employs the set of template rules that
was learned from examples to overcome the problem of
computational complexity.
2.4.4
      </p>
      <sec id="sec-8-1">
        <title>Log-Linear Models</title>
        <p>Log-Linear models are generally used in Natural Language
processing. The flexibility of this model is its major benefit; it
allows the use of rich set of features.</p>
        <p>
          In [
          <xref ref-type="bibr" rid="ref13">14</xref>
          ], log-linear models were used to bring to null the
assumption that existing systems were feature independent.
Consequently, it was also showeshown empirically that using
log-linear models produced better extracts than naïve-bayes
model. The conditional log-linear model used by the author can be
stated as follow:
…………..(2)
Let c = label, s = item we want to label, fi = i-th feature, λi = the
corresponding feature weight and Z (s) = ∑c exp (∑i λi fi (c,s)).
        </p>
      </sec>
    </sec>
    <sec id="sec-9">
      <title>3.4 Neural Networks and Third Party Features</title>
      <p>
        The automatic text summarization system developed in [
        <xref ref-type="bibr" rid="ref14">15</xref>
        ] had
learning ability. This was done through combination of a
statistical approach, extraction of keywords, neural network and
unsupervised learning. The process used involved three steps: step
one involved removal of stop words like “a” and stemming which
is done by removing suffixes and prefixes to convert a word to its
stem. Step two involves keywords are extracted by computing the
matrix of the term frequency against the inverse document
frequency, the most frequent terms listed are the keywords to be
extracted for the summary. For the final step, the model checks
for stop words again to be sure that no stop word is selected as
keyword after which it selects sentences containing keywords to
be added to the summary.
      </p>
      <p>
        NetSum [
        <xref ref-type="bibr" rid="ref15">16</xref>
        ] was the first to use neural network ranking algorithm
and third-party datasets for automatic text summarization. The
authors trained a system that learned from a train set containing
labels of best sentences from which features are extracted. From
the train set, the system learns how features are distributed in the
best sentences and it gives a result of ranked sentences for each
document, the ranking is done using RAnkNet [
        <xref ref-type="bibr" rid="ref16">17</xref>
        ]
      </p>
    </sec>
    <sec id="sec-10">
      <title>3. SENTENCE SCORING AND SENTENCE</title>
    </sec>
    <sec id="sec-11">
      <title>REDUCTION MODELS</title>
      <p>Sentence score is a value that determines the sentences that are
relevant to the input text. As shown in Figure 1, in our
architecture, the input to the system is a single document.
Sentence scoring occurs at the first stage; significant sentences are
identified and extracted. The second stage involves the sentence
reduction module; the extracted sentences from the sentence
scoring module are processed, grammar checking and removal of
target structures is done.</p>
    </sec>
    <sec id="sec-12">
      <title>3.1 Sentence Scoring Module</title>
      <p>User Input
Extracted</p>
      <p>Sentences
Sentence
Extraction


</p>
      <p>Sentence
resemblance to
query
Cue Phrases
Word frequency
Score =</p>
      <p>
Score =

Sentence resemblance to query: This is modelled after
sentence resemblance to title which calculates a score based
on the similarity between a sentence and the title of a
document. So sentence resemblance to query calculates a
score based on the similarity between a sentence and the
user query which means that any sentence that is similar to
the query or includes words in the query are considered
important. And the score will be calculated using the
following formula:
----(1)
Where nQW = number of words in query</p>
      <p>Cue Phrases: the justification of using this feature is
that the presence of some words likes “significantly”,
“Since” point to important gist in a document and a
score is assigned to such sentences. The score is
computed using:</p>
      <p>
        ----------(2)
Word frequency, is a useful measurement of significance
because it is revealed in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] that an author tend to repeat
certain words when trying to get a point across. So sentences
that contain frequently occurring words are considered to be
significant. The algorithm involves:
      </p>
      <p>II.</p>
      <sec id="sec-12-1">
        <title>Breaking sentence into tokens</title>
      </sec>
      <sec id="sec-12-2">
        <title>For each token, if the token already exists in array, syntactic parsing, grammar checking, removal of target structures and display of output. Increment its count,</title>
        <p>Else add token to array and</p>
        <p>Set initial count to 1.</p>
        <p>The boolean formula below is used to decide the sentences to be
selected for further processing:
(SrqScore &gt;= 0.5 || (CpScore &gt;= 0.1 &amp;&amp; WfScore &gt;=3) ---(3)
Where SrqScore is Sentence resemblance to query Score, CpScore
is Cue phrase score and WfScore is Word Frequency score.
3.2</p>
      </sec>
    </sec>
    <sec id="sec-13">
      <title>Sentence Reduction Module</title>
      <p>
        In the sentence scoring module, the original document and the
extracted sentences from the sentence scoring module is processed
so as to remove irrelevant phrases from the document to make the
summary concise, the sentence reduction algorithm is described in
details in [
        <xref ref-type="bibr" rid="ref17">18</xref>
        ]. The processing involves:
      </p>
      <sec id="sec-13-1">
        <title>Syntactic Parsing</title>
        <p>Stanford parser, a syntactic parser is used to analyze the structure
of the sentences in the document and a sentence parse tree is
produced. The other stages involved in the sentence reduction
module add more information to the parse tree, these information
aids the final decision to be made.</p>
      </sec>
      <sec id="sec-13-2">
        <title>Grammar checking</title>
        <p>We go back and forth on the sentence parse tree, node by node to
identify parts of the sentence are important and must not be
removed to make the sentence grammatically correct. For
example, in a sentence, the main verb, the subject, and the
object(s) are essential if they exist.</p>
      </sec>
      <sec id="sec-13-3">
        <title>Removal of Target Structures</title>
        <p>For this research work we will be using the main clause algorithm
for sentence reduction. In this algorithm, the main clause of a
sentence is obtained and in that main clause we identify the target
structures which are the structures to be removed and they are
adjectives, adverbs, appositions, parenthetical phrase, and relative
clauses. A reduced sentence is gotten after the targeted structures
has been identified and removed from the sentence parse tree.</p>
      </sec>
      <sec id="sec-13-4">
        <title>Summary Construction</title>
        <p>We once again go back and forth on the sentence parse tree and
see if the reduced sentences are grammatically correct. The
reduced sentences are then merged together to give the final
summary.</p>
        <p>After the sentence reduction module carries out all four steps, a
concise and coherent summary is expected as output.</p>
      </sec>
    </sec>
    <sec id="sec-14">
      <title>4. IMPLEMENTATION</title>
      <p>The system allows a user to input a document; which is then
prepossessed and the sentences ranked. The high ranked sentences
are then extracted for further processing. The result is then viewed
by the user. As shown in figure 3, the flow of activities includes
uploading of input file, preprocessing, assignment of scores,</p>
    </sec>
    <sec id="sec-15">
      <title>4.1 Implementation Resources</title>
      <p>The following resources were used for the implementation of our
summarization system:


</p>
      <p>Java: it was considered a good choice for developing
the summarization system because it makes more
efficient use of memory needed for a system of this
nature, it is faster and more effective, and also provides
more data structures to handle the different data types
used in the implementation of the proposed algorithm.
Stanford Parser: The Stanford parser was used as the
tagging tool in the sentence reduction module of our
implementation. The Stanford parser analyses the
sentences and provides us with the parts of speech of the
words in a sentence, as well as the class e.g. adverbial
phrase, adjectival phrases, etc., different parts of the
sentence belongs to.</p>
      <sec id="sec-15-1">
        <title>Gate (General Architecture for Text Engineering):</title>
        <p>We used GATE in the development our algorithm to
integrate the Stanford parser and its other modules such
as the sentence splitter and tokenizer which properly
handles complexities that may occur in long articles
than ordinary tokenizers cannot handle properly. The
integration of the parser and this other modules was
used to create an application pipeline which was then
used as a plugin in the implementation and development
of our summarisation system.</p>
      </sec>
    </sec>
    <sec id="sec-16">
      <title>5. RESULTS AND DISCUSSION</title>
      <p>To test our summarisation system, we obtained random articles
online from Wikipedia to use as our input document. The article is
then saved in a text (.txt) file to be used in our system. Wikipedia
references are disregarded during our extraction and not included
in the content of the article to use as input document for our
summarisation system, as they do not provide any value of
importance to the overall article and the summary we want to
generate. For this task, we selected an article about Web 2.0,
however, it is important to note that any article could be selected
and used.
We evaluated our summarization system using 3 standard
parameters; (1) precision (the fraction of the returned summary
that is relevant to the query posed by the user); (2) recall (the
fraction of the relevant information contained in the document
that was returned by the summarization system); and (3)
Fmeasure (the weighted harmonic mean of precision and recall).
We had a human subject read two articles, one was the web 2.0
article and the other was a Blackberry passport review.
The web 2.0 article had 179 sentences, the summary presented by
the subject contained 110 sentences and our system produced a
summary containing 112 sentences. 92 sentences were present in
both the summary made by the subject and the summary made by
our system.</p>
      <p>The second document, a blackberry passport review contained
129 sentences. In the summary presented by the subject we had 40
sentences and our summarizers’ summary, we had 57 sentences.
27 sentences were present in both the summary made by the
subject and the summary made by our system. Based on
comparison with the well-known Copernic summarizer which
produces summary based on statistical and linguistic methods and
the FreeSummarizer that produces summary based on specified
number of sentences, our system gave high recall values of 83.6%
and 67.5% respectively; an indication that the sentences in our
system’s summary contain more relevant information such that
selected sentences are relevant to the query pos. ed by the user.
In conclusion, this work can be extended to multi-document
summarization whereby the source text is more than one; the
resulting summary may contain a larger amount of information
when compared to a single document summarization. Also, it can
be extended to other languages apart from English language.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Wells</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          (
          <year>2009</year>
          ).
          <article-title>Advantages of automatic text summarization</article-title>
          , http://ezinearticles.com/?3-
          <string-name>
            <surname>Advantages-</surname>
          </string-name>
          ofAutomatic-
          <source>Text-Summarization&amp;id=3465270 downloaded on 11 September</source>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Luhn</surname>
            ,
            <given-names>H. P.</given-names>
          </string-name>
          (
          <year>1958</year>
          ).
          <article-title>The automatic creation of literature abstracts</article-title>
          ,
          <source>IBM Journal of research and development</source>
          , Vol.
          <volume>2</volume>
          , No.
          <issue>2</issue>
          , pp.
          <fpage>159</fpage>
          -
          <lpage>165</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Baxendale</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          (
          <year>1958</year>
          ).
          <article-title>Machine-made index for technical literature - an experiment</article-title>
          .
          <source>IBM Journal of Research Development</source>
          ,
          <volume>2</volume>
          (
          <issue>4</issue>
          ):
          <fpage>354</fpage>
          -
          <lpage>361</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Jing</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          (
          <year>2000</year>
          ).
          <article-title>Sentence Reduction for Automatic Text Summarization</article-title>
          ,
          <source>In Proceedings of the Sixth Applied Natural Language Processing Conference</source>
          , pp.
          <fpage>310</fpage>
          -
          <lpage>315</lpage>
          . Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Jing</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <article-title>and</article-title>
          <string-name>
            <surname>McKeown</surname>
            ,
            <given-names>K.R.</given-names>
          </string-name>
          (
          <year>2000</year>
          ).
          <article-title>Cut and Paste Based Text Summarization</article-title>
          ,
          <source>In Proceedings of the 1st Meeting of the North American Chapter of the Association for Computational Linguistics</source>
          , pp.
          <fpage>178</fpage>
          -
          <lpage>185</lpage>
          . Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Kupiec</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pedersen</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          (
          <year>1995</year>
          ).
          <article-title>A trainable document summarizer</article-title>
          ,
          <source>In Proceedings of the 18th annual international ACM SIGIR conference on Research and development in information retrieval</source>
          , pp.
          <fpage>68</fpage>
          -
          <lpage>73</lpage>
          . ACM.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Larsen</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          (
          <year>1999</year>
          ).
          <article-title>A trainable summarizer with knowledge acquired from robust NLP techniques</article-title>
          .
          <source>Advances in Automatic Text Summarization</source>
          , pp.
          <fpage>71</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>C. Y.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Hovy</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          (
          <year>1997</year>
          ).
          <article-title>Identifying topics by position</article-title>
          ,
          <source>In Proceedings of the fifth conference on Applied natural language processing</source>
          , pp.
          <fpage>283</fpage>
          -
          <lpage>290</lpage>
          . Association for Computational Linguistics
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Le</surname>
            ,
            <given-names>N. M.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Horiguchi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          (
          <year>2003</year>
          ).
          <article-title>A new sentence reduction based on decision tree model</article-title>
          ,
          <source>In Proceedings of the 17th Pacific Asia Conference on Language, Information and Computation</source>
          , pp.
          <fpage>290</fpage>
          -
          <lpage>297</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Knight</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Marcu</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          (
          <year>2002</year>
          ).
          <article-title>Summarization beyond sentence extraction: A probabilistic approach to sentence compression</article-title>
          ,
          <source>Artificial Intelligence</source>
          , Vol.
          <volume>139</volume>
          , No.
          <issue>1</issue>
          , pp.
          <fpage>91</fpage>
          -
          <lpage>107</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Ghahramani</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          (
          <year>2001</year>
          ).
          <article-title>An Introduction to Hidden Markov Models and Bayesian Networks</article-title>
          ,
          <source>International Journal of Pattern Recognition and Artificial Intelligence</source>
          , Vol
          <volume>15</volume>
          , No.
          <issue>1</issue>
          , pp.
          <fpage>9</fpage>
          -
          <lpage>42</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [13]
          <string-name>
            <surname>Nguyen</surname>
            ,
            <given-names>M. L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Horiguchi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shimazu</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Ho</surname>
            ,
            <given-names>B. T.</given-names>
          </string-name>
          (
          <year>2004</year>
          ).
          <article-title>Example-based sentence reduction using the hidden markov model</article-title>
          ,
          <source>ACM Transactions on Asian Language Information Processing (TALIP)</source>
          , Vol.
          <volume>3</volume>
          , No.
          <issue>2</issue>
          , pp.
          <fpage>146</fpage>
          -
          <lpage>158</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [14]
          <string-name>
            <surname>Osborne</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          (
          <year>2002</year>
          ).
          <article-title>Using maximum entropy for sentence extraction</article-title>
          ,
          <source>In Proceedings of the ACL-02 Workshop on Automatic Summarization</source>
          , Vol.
          <volume>4</volume>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          . Association for Computational Linguistics.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [15]
          <string-name>
            <surname>Yong</surname>
            ,
            <given-names>S. P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Abidin</surname>
            ,
            <given-names>A. I.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>Y. Y.</given-names>
          </string-name>
          (
          <year>2005</year>
          ).
          <article-title>A Neural Based Text Summarization System</article-title>
          ,
          <source>In Proceedings of the 6th International Conference of DATA MINING</source>
          , pp.
          <fpage>45</fpage>
          -
          <lpage>50</lpage>
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [16]
          <string-name>
            <surname>Svore</surname>
            ,
            <given-names>K. M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Vanderwende</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Burges</surname>
            ,
            <given-names>C. J.</given-names>
          </string-name>
          (
          <year>2007</year>
          ).
          <article-title>Enhancing Single-Document Summarization by Combining RankNet</article-title>
          and
          <string-name>
            <surname>Third-Party</surname>
            <given-names>Sources</given-names>
          </string-name>
          , In EMNLP-CoNLL, pp
          <fpage>448</fpage>
          -
          <lpage>457</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Burges</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shaked</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Renshaw</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lazier</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Deeds</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hamilton</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          , &amp;
          <string-name>
            <surname>Hullender</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          (
          <year>2005</year>
          ).
          <article-title>Learning to rank using gradient descent</article-title>
          .
          <source>In Proceedings of the 22nd international conference on Machine learning</source>
          (pp.
          <fpage>89</fpage>
          -
          <lpage>96</lpage>
          ). ACM.
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Silveira</surname>
            ,
            <given-names>S. B.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Branco</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          (
          <year>2014</year>
          ).
          <article-title>Sentence Reduction Algorithms to Improve Multi-document</article-title>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>