<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Context-Aware Model of Abstractive Text Summarization for Research Articles</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Gopinath Dineshnath</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Selvaraj.Saraswathi</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science Engineering, Pondicherry Engineering College</institution>
          ,
          <addr-line>Pillaichavady, Puducherry, 605014</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Department of Information Technology, Pondicherry Engineering College</institution>
          ,
          <addr-line>Pillaichavady, Puducherry, 605014</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>ive text summary and outperforms at sentence level Rouge-L measures 9.32 and summary level measures 89.65.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Modern days, internet becomes the integral part
of human and act as information highway. The
primary source of information in digital world
is Internet and it is boon for academicians,
bloggers, students and researcher fraternity.
Information available in Internet comprises of
massive flow of information, which makes
retrieval process complex with respect to
context-specific content. Scientific article
prevailing now with ocean of research domains
makes difficult to scholar cope-up, grasp and
streamline documents relevant to their interest.
Query based search [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] for specific domain also
fetch many relevant articles that is difficult task
to categorize surpass human processing
capabilities. In such scenario, automatic text
summarization of articles is fruitful solution in
terms of reducing time effort for reviewing
entire articles and grab gist of information
enclosed in it. Basically, summaries generation
in two ways; single-document summaries
produce a summary from a given single source
and multi-document summaries in which
different but related documents are summarized
by comprises only the vital materials or main
ideas in a document in less space.
      </p>
      <p>
        There is a vast difference between automatic
multi-document summarization of generic texts
to that of scientific articles. The major
difference [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] between generic-text and research
article summarization is; research article
consists of different section namely abstract,
introduction, literature survey, methods, results
and discussions, whereas generic text’s scope is
extracted from first few sentences in first
paragraphs and entire section holds at
maximum 500 words.
      </p>
      <p>In general, abstract and citation texts in
scientific articles are considered for automatic
summarization system.</p>
      <p>
        Abstract section[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] is biased to author
findings, author’s own contribution, and
evaluation metrics. In simpler way, abstract
outlines the domain and list of findings in crisp
manner depends upon the type of articles
(review/original).
      </p>
      <p>
        Citation sensitive or citation-based summaries
[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] is another type of Scientific article
summarization, major task in summary
production is clear-cut distinction of cited and
non-cited text is performed. Citation summary
system performs categorization of every
sentence and labeled to citation or non-citation
one. Later, evaluation measures based on
similarity between each sentence in the
reference article and citation sentences and then
grouped it into one of two classes: cited or
noncited. Abstractive Multi document
summarization [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] selectively picks up either
first sentence of abstract or introduction of a
paper since it is comprised of background
information of research topic. Construction of
appropriate title for article involves
interpretation and integration of the concepts
from multiple sentences of the abstract. Apart
from that there exist multiple challenging issues
like content organization, sentence
compression and fusion, and paraphrasing
sentences.
      </p>
      <p>
        All summarization system should meet the
summary length constraints as and other
parameters specified by the user or
summarization system is known as controllable
summarization. Controllable summarization [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]
is the main criteria in summarization system
which specifies the length of summary
generation in accordance to the entities on
which it focuses on and mimics the source’s
style. User may define the high-level attributes
for summary generation. Summary generation
is controlled by specific control variables,
length, and source style, entities of interest and
summarizing only remaining portions of the
document. For instance, blog summarization,
the primary thing is to derive representative
words from comments and then selection of
paramount sentences from the blog post which
consists of representative words.
      </p>
      <p>Context aware components are usually
meant task of inferring contextual information.
Contextual information detection might be
detection ranging from topic community,
paragraphs analysis, sentences and words by
statistical computation measures. The most
well-known computation is Set-Calculus
techniques to find sentence similarity via
Union, Jaccard Co-efficient, Cosine -Similarity
measures followed by normalization
techniques. Our main focus is Context aware
information inference from multiple
documents.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Literature review</title>
      <p>
        Lloret et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] have applied both extractive and
abstractive summarization procedures for
scientific article abstracts. The extractive
summarizer (compendium E) is developed to
perform conventional preprocessing such as
breaking sentences, assigning tokens,
stemming, lemmatization and PNG markers,
tagging and removing duplicates at various
sentence levels. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]A mixture of both extractive
and abstractive technique (compendium E−A)
is developed to support compendium E as base
to incorporates sorted information which are
relevant. Relevancy identification with respect
to every sentences, assigns a score that
emphasize its importance based on code
quantity principle (CQP)[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] compendium E−A
derives abstractive summary by utilizing top
ranked sentences with chronological ordering.
      </p>
      <p>
        Saggion [11] utilized pretrain models
for learning and transformations for the
problem of abstract generation. The initial
summary generation from abstracts are
generated and transformed to model based
learning. The learning models assists with
examples from corpus. Further, abstracts are
gathered from GATE [12]and Weka [13]
environment. Abstractive text summarization
also known as natural language generation in
natural language processing Paraphrasing of
sentences is also another important criterion in
natural language generation Paraphrasing of
sentences [14] involves substitution of relevant
verbatim and modifying the direct to indirect
speech or vice versa. The vector representation
is purely focus on various sources of features
namely LDA, D2V, W2V and encoding
schemes[15]..LDA[16] explores semantic
associations, D2V vectors finds contextual
word vectors along with documents. Contextual
aware model phase is concerned with
contextual theme and dependency phrase
extraction from multi documents using directed
graph. minimal spanning tree is constructed for
edges algorithm using Chu Liu Edmonds.
(CLE)[
        <xref ref-type="bibr" rid="ref10">17</xref>
        ] Knowledge base or ontology- based
context aware model useful for domain
classification and annotation.[
        <xref ref-type="bibr" rid="ref11">18</xref>
        ]Evaluation of
document summarization, Document
Understanding conference (DUC)[
        <xref ref-type="bibr" rid="ref12">19</xref>
        ]
benchmark datasets is used generally. Various
datasets such as TAC[
        <xref ref-type="bibr" rid="ref13">20</xref>
        ],shared task for text
processing and document summarization
Similarly, DPIL[
        <xref ref-type="bibr" rid="ref14">21</xref>
        ] for paraphrasing Indian
languages for text summarization.
      </p>
      <p>W(Vi) = (1 – d) + d ∗</p>
      <p>overlap(si ,sj )
∑Vj∈ln(Vi) ∑ Vk∈Out(Vj )overlap(sj ,sk) W(Vj)
(1)
.</p>
      <p>The edge weights for nodes are assigned
using status score on the basis of inwards and
outwards edges, W(Vi) represents status score
assigned to vertex Vi. In (Vi) and Out (Vj) are
inwards and outwards edges, points from
particular node. After several iterations, each
sentence in the document is assigned with a
score. The top-n sentences are selected and
ranked which constructs the summary for the
document. There exist some dependent phrases
in graph. The dependent phrases are cyclic in
nature and some disjoint nodes. Such, disjoint
nodes in graph are connected using CLE
algorithm. On basis of lexical contexts,
proposed procedure performs well than
traditional keyword-based algorithms. We
enhance the extractive summary production by
adding co-occurrence measures to ensure
Concept-based Point wise Mutual Information.
(CPMI) CPMI weighs the different section in
paragraphs gradually it weights decreases from
beginning of paragraph to end of paragraph in
document. CPMI measures support
distributional semantics among phrases. CPMI
weightage scheme H(ƥ) is expressed in
equation 2.</p>
      <p>H(ƥ)
=
{</p>
      <p>C − positiveconstantƥ ∗ B,
if ƥ &lt; − (llooggBC)
1
otherwise
}
,
(2)</p>
    </sec>
    <sec id="sec-3">
      <title>3. Context Aware Model Phase</title>
      <p>Context aware component model proposed
follows contextual keyword interpretation,
topic detection and topic clusters formation.
Context aware component model also
determines vector manipulation by bag of
words and skip gram key terms with respect to
specific documents. One hot encoding scheme
is used to slide over word vectors for prediction
of context vectors. Skip gram models skips the
selection of common words rather than
contextual words. N gram models hold
predefined size ‘n’ that triggers the selection of
contextual words upon size limited to ‘n’. Both
skip gram and N gram models are desirable
notion to pick keywords in context based on
their lexicography collocations. The Proposed
diagram is given in figure-1.
Contextual matching phrase from multi
documents is proposed to retrieve thematic
portions with similar sentence phrases
sequentially from one document to another. The
graph based contextual word is intuitive way to
Formula for Extraction of Eq</p>
      <p>Features. n.o:
 = NuamvebreargoefletnitgltehF(eTaittlue)res (3)</p>
      <p>Tag sum = 11−+ eeα–α (4)</p>
      <p>∑L(Si) (5)
(Sbi) = Lmax</p>
      <p>Dsm( ) = T W( ) (6)
represent sentences as node and corresponding
contextual word as vertices. The path projecting
from node to node via outcoming and incoming
vertices provides a notion either matching
target phrases or discriminating phrases.
subsequent dependent phrases need to be
included in directed graph. The procedure for
contextual theme and dependency phrases
extraction is shown below.</p>
      <p>Step 1 Accept text and store it in text buffer.
Step 2 new word falls into below categories.
Step 2 a): If first word then adds to graph G.
Step 2 b): If fresh word then appends to G.
Step 3 Go to step 2 until words overlap.
Step 4 If overlap, status score using (1).
Step 5 Extract similar texts and update in G.
Step 6 Construct Digraph using CLE.</p>
      <p>Step 7 Do updates to infer adjacent edges.</p>
      <p>Step 8: Output the phrases.
3.2.</p>
    </sec>
    <sec id="sec-4">
      <title>Feature Extraction</title>
      <p>The feature extraction for document
summarization includes title feature, proper
nouns, Named Entities Recognition (NER) and
parts of speech tagging, sentence boundary
analysis and distributional semantic analysis.
Title feature scoring scheme is based on the
ratio of mean number of titles present to that of
average length of title. The formula for title
feature is expressed in equation. (3). The proper
nouns are generally recognized as title words
and minimum number of words to accept as
title. NER marks or labels the salient sentences
which is considered for summary. The scoring
scheme is expressed in equation. (4). where α =
(t(s) − µ) /σ (sigmoid function) aggregates
mean count of regular expressions, case level
and numeric literals. Sentence boundary
calculation is expressed in equation (5).
Distributional Semantic Analysis weighs
thematic concepts to find word co-occurrence.
It is scored using formula. (6). Column one
represents features, similarly column two
represents formulae used for computation and
column three indicates optimal features
customized to produce extractive to abstractive
summary.in Table-1. The optimal features and
their convergence in Adaptive Glowworm
Optimization (AGWO) is discussed in section
3.4.</p>
      <p>Table 1 Optimal Features Extraction.</p>
      <sec id="sec-4-1">
        <title>Features</title>
      </sec>
      <sec id="sec-4-2">
        <title>Title feature</title>
      </sec>
      <sec id="sec-4-3">
        <title>Named entity</title>
        <p>recognition and</p>
      </sec>
      <sec id="sec-4-4">
        <title>Tagging</title>
      </sec>
      <sec id="sec-4-5">
        <title>Sentence boundary</title>
      </sec>
      <sec id="sec-4-6">
        <title>Distributional</title>
      </sec>
      <sec id="sec-4-7">
        <title>Semantic Analysis</title>
      </sec>
      <sec id="sec-4-8">
        <title>Glowworm</title>
        <p>optimization</p>
      </sec>
      <sec id="sec-4-9">
        <title>Luciferin update phase</title>
      </sec>
      <sec id="sec-4-10">
        <title>Movement phase</title>
      </sec>
      <sec id="sec-4-11">
        <title>Neighborhood Phase</title>
      </sec>
      <sec id="sec-4-12">
        <title>Words and conceptual level</title>
        <p>3.3.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Vector Formulation</title>
      <p>Word embeddings are feature vectors that
represent words holds the property that similar
words have similar feature vectors. The
question might rise in mind where the
embeddings come from. The response is
(again): they are learned from data. Several
algorithms exist to learn word embeddings. We
consider only one of them: word2vec, and sole
version of word2vec called skip-gram, which is
well-known and currently utilized in practice.
Word embedding learning, our goal is to build
a model which we can use to convert a one-hot
encoding of a word into a word embedding. Let
our dictionary contain 10,000 words or
Gigaword Corpus. Skip-gram model performs for
given sentence, selection of a word is feed into
classifier, and predict words before and after the
selected word in a fixed window. Negative
sampling provides better vectors for frequent
words with low dimension.</p>
      <p>Latent Dirichlet (LD) allocation is a
possibility-based mechanism viable for
assortments, for example, text assortments. LD
consolidates the documents as a blend of shifted
topics; every unit involves words that have a
spun affiliation that exists between them. Also,
word choice simply dependent on the numerical
idea of likelihood. Recursively determining the
interaction of themes and words is done for the
phase of a lonely record or a large number of
documents. At long last, yields the record
which comprises of different subjects.</p>
      <p>LD allocation algorithm performs the
following:</p>
      <p>1) Determine the number N of words in the
document concerning probability distribution is
Poisson.</p>
      <p>2) Pick a merge of focuses for the report
from a predefined set of K subjects as
demonstrated by the Dirichlet movement.</p>
      <p>3) Produce individually word in the list of
terminology as follows language Vocabulary
(V).</p>
      <p>a) Choose a subject;
b) Choose a word in this subject.
3.4.</p>
    </sec>
    <sec id="sec-6">
      <title>Sentence Ranking</title>
      <p>
        Sentence ranking phase, chiefly performs
identification of prominent sentences with
pertinent information and free from
redundancy. It selects topmost sentences from
documents and produces summary with
application of traditional maximization
algorithm like EM.[
        <xref ref-type="bibr" rid="ref16">23</xref>
        ]. finally produces
extractive summary. However, Extractive
summary lacks positional placement of
sentences. hence there is a need to revisit
sentence positions.[
        <xref ref-type="bibr" rid="ref17">24</xref>
        ].
3.5.
      </p>
    </sec>
    <sec id="sec-7">
      <title>Extractive Summary</title>
      <p>Extractive summary is created with top n
sentences for research articles summarization
Latter summary is transformed to decision
matrix. Decision Matrix will keep track of
extractive summary to make compatible for
abstractive text summary with usage measure
and penalty measure. Hence, optimization
algorithm is used to remove exact replica of
original text produced in extractive summary.
Co-reference resolution is also handled in
abstractive text summary generation.
Meanwhile construction of cosine similarity
was rapid, useful, and seems reasonable.
3.6.</p>
    </sec>
    <sec id="sec-8">
      <title>Optimization</title>
      <p>li(t + 1) = (1 −  )li(t) + ʋJ (xi (t + 1));
Ni(t) = {j: dij &lt; rid(t); li (t) &lt; lj (t)}
   ( + 1) = 
{  ,  {0,
   ( ) +  (ǹ
− |  ( )|}};
(7)
(8)
(9)</p>
      <p>
        Glowworm Optimization (GO)[
        <xref ref-type="bibr" rid="ref15">22</xref>
        ]
comprises of three phases namely Luciferin
update phase, Neighborhood phase and
movement phase. Adaptive Glowworm
Optimization (AGO) is proposed for
tailormade features to acquire vectors or extract
features to frame the summary. The
optimization principle is based on five features
and their application phase is listed in Table-1.
      </p>
      <p>Sentence position is additional feature to
revisit sentence with appropriate ordering.
Positioning of sentences which is most vital
part in the summary generation have higher
weights. The feature associated with sentence
length; hence we have minimal set of 25 words
to accept as a sentence. F5 feature is in
movement phase of Glow worm optimization
with lucerifin value or luminous quotient, affine
towards the similar topics.</p>
      <p>Luciferin update phase, sentences are
concatenated with respect to the relevancy.
Relevancy is determined by feature with
respect to title and all sentences in document.
Luciferin update phase, movement phase and
neighbor phase are expressed as equation in (7)
(8) and (9) respectively Luciferin
enhancement(ʋ) depends upon Proper Nouns.</p>
      <p>J(xi(t))-objective function which maximize
weights of every proper nouns. Luciferin decay
constant gradually decreases when common
noun exists.</p>
      <p>Movement Phase, forms local clusters based
on decision range. Sentences are of similar
contexts likely to move based on entity-tagging
features. Finally, neighbor Phase performs
chronological sorting of clustered sentences to
produces summary.</p>
    </sec>
    <sec id="sec-9">
      <title>4. Results and Discussions</title>
      <p>The extractive summarization shows better
results with proposed procedure, the summary
produced purely relies on lexical features and
surpass traditional keyword ranking schemes.
yellow color denotes dependency phrases and
green color denotes contextual theme. The
output of context aware component of
extractive summarization is shown as well as
output of abstractive text summarization is also
shown.</p>
    </sec>
    <sec id="sec-10">
      <title>Training and Testing</title>
      <p>For training purpose, Document understanding
Conference (DUC) data set taken into
considerations. The precise explanation of
DUC data sets and DUC data is customized,
which is free from least significant words or
stop words according to port-stemmer’s
algorithm. Recall oriented understudy Gist
evaluation, Rouge(R) is also considered for
evaluation. R falls into many variants like
Runigram, R-bigram, R-Longest common
Subsequence and R-N gram classes.</p>
      <p>Multiple documents of artificial intelligence
domain for testing and performed various
measures like R-1, R-2, and R-L (Longest
Common Subsequence) scoring. At
Sentencelevel, computes longest common subsequence
(LCS) between two pieces of text ignores new
lines and summary-level, newlines in the text
are interpreted as sentence boundaries, and the
LCS is computed between each pair of
reference and candidate sentences, and their
results are tabulated below in table-2.</p>
      <p>
        The proposed AGO performs well in sentence
and summary level than traditional methods.
similarly, contextual theme detection also out
performs than traditional schemes like lexrank,
maximum relevance, loglikelihood ranking
[
        <xref ref-type="bibr" rid="ref18">25</xref>
        ],[
        <xref ref-type="bibr" rid="ref19">26</xref>
        ],[
        <xref ref-type="bibr" rid="ref20">27</xref>
        ] and other centrality measures as
baseline evaluation.
0.4611 0.1342
0.3716 0.0757
      </p>
      <p>Extractive Summary</p>
      <p>A large number of methods applied in
the field of extractive summarization
over the years. Scoring sentences for
such summary is tedious task. Many
researchers putting so much effort to
improve the quality of summary.
Document summarization focus both
quality and coverage of content.
Clustering of sentences in document
summarization shown promising results
to discover topics. A Fuzzy oriented
clustering for summarization of
multidocuments. Compendium- a summarizer
tool, generates relevant summary free
from redundancy Collabsum clustering
process both inter and intra document
relationship and forms clusters. Clusters
in turn apply graph based ranking
methodology. FEOM…genetic
algorithm…graph -based
approach…probabilistic model.</p>
      <p>Various types of Sentence clustering
techniques applied to document
summarization Sentence scoring, topic
coverage, relevant sentences and
summarization quality are main
components in summary production.
Clustering algorithms for sentence
scoring and grouping similar sentences
according to topics conveyed in
document. A fuzzy based, evolutionary
based clustering also successfully
applied in conjunction with other
graphbased approaches to provide summary.</p>
    </sec>
    <sec id="sec-11">
      <title>5. Conclusion</title>
      <p>Abstractive text summarization for research
articles generates sentences individually using
glowworm optimization with six associated
features. In addition, decision matrix with
elitism identification is formulated to choose
summary sentences from both extractive
summary sentences and abstractive summary
sentences with consistency as necessary
condition. Extractive summary is reduced to
more than 80% to generate abstractive
summary. Extractive summary with 661 word
tokens is produced as output in first phase.
Later, decision matrix with Elitism
identification produces abstractive summary
with 84 tokens is obtained as final output.
Proposed multi-document directed graph
contextual matching phrases, Rouge-L
measures in sentence level is 12.08 and
RougeL measures in summary level is 83.95 for
extractive summary. Similarly, Rouge-L
measures in sentence level is 9.32 and Rouge-L
measures in summary level is 89.68 for
abstractive summary. A novel model has been
implemented to be ample enough to provide
multi objectives and to convince the
instantaneous needs. Ultimately, this study will
inspire many researchers to further explore and
apply the various types of Swarm intelligence
while solving the summarization tasks,
specifically in the abstractive text
summarization (ATS) field.</p>
    </sec>
    <sec id="sec-12">
      <title>6. Future works</title>
      <p>
        Decision matrix performs combination of
Sentences from extractive summary are
assessed and deemed to be fit for abstractive
summary are analyzed in conjunction with
input from optimization algorithm with
associated six features. Selection of best
sentences and worst sentences based on their
usage and penalty is awarded to compose
summary. Global decision matrix performs
elitism identification (algorithm) and outputs
sentences with sentence flow as criterion.
However, decision matrix follows Analytical
Hierarchical Processing (AHP) [
        <xref ref-type="bibr" rid="ref21">28</xref>
        ] with user
defined decision values and their decisions are
normalized. We can extend the normalized
vectors by using fuzzy [
        <xref ref-type="bibr" rid="ref22">29</xref>
        ] based membership
assessment as stated by Charugupta,et.al.[
        <xref ref-type="bibr" rid="ref23">30</xref>
        ]
      </p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Shafiei</given-names>
            <surname>Bavani</surname>
          </string-name>
          , Elaheh, Mohammad Ebrahimi, Raymond Wong, and
          <string-name>
            <given-names>Fang</given-names>
            <surname>Chen</surname>
          </string-name>
          .
          <article-title>"A querybased summarization service from multiple news sources."</article-title>
          <source>In 2016 IEEE International Conference on Services Computing (SCC)</source>
          , IEEE, (
          <year>2016</year>
          ): pp.
          <fpage>42</fpage>
          -
          <lpage>49</lpage>
          . doi:
          <volume>10</volume>
          .1109/SCC.
          <year>2016</year>
          .
          <volume>13</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Bharti</surname>
            ,
            <given-names>Santosh</given-names>
          </string-name>
          <string-name>
            <surname>Kumar</surname>
            , Korra Sathya Babu, Anima Pradhan,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Devi</surname>
            ,
            <given-names>T. E.</given-names>
          </string-name>
          <string-name>
            <surname>Priya</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Orhorhoro</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <string-name>
            <surname>Orhorhoro</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          <string-name>
            <surname>Atumah</surname>
            , E. Baruah, and
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Konwar</surname>
          </string-name>
          .
          <article-title>"Automatic keyword extraction for text summarization in multi-document e-newspapers articles."</article-title>
          <source>European Journal of Advances in Engineering and Technology</source>
          (
          <year>2017</year>
          ): 4, pp.
          <fpage>410</fpage>
          -
          <lpage>427</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Chowdhury</surname>
            ,
            <given-names>S. M.</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>and Mazharul</given-names>
            <surname>Hoque</surname>
          </string-name>
          .
          <article-title>"A Review Paper on Comparison of Different Algorithm Used in Text Summarization." Intelligent Data Communication Technologies and Internet of Things: ICICI 2019 38 (</article-title>
          <year>2019</year>
          ):
          <fpage>114</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Cohan</surname>
            , Arman, and
            <given-names>Nazli</given-names>
          </string-name>
          <string-name>
            <surname>Goharian</surname>
          </string-name>
          .
          <article-title>"Scientific article summarization using citationcontext and article's discourse structure</article-title>
          .
          <source>"</source>
          (
          <year>2017</year>
          ).
          <source>arXiv preprint arXiv:1704.06619</source>
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Nouf</given-names>
            <surname>Ibrahim</surname>
          </string-name>
          <string-name>
            <surname>Altmami</surname>
          </string-name>
          , Mohamed El Bachir Menai, “
          <article-title>Automatic summarization of scientific articles: A survey</article-title>
          .
          <source>” Journal of King</source>
          Saud University-Computer and Information Sciences, (
          <year>2020</year>
          )
          <article-title>doi</article-title>
          .:
          <volume>10</volume>
          .1016/j.jksuci.
          <year>2020</year>
          .
          <volume>04</volume>
          .020.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Fan</surname>
            , Angela, David Grangier,
            <given-names>and Michael</given-names>
          </string-name>
          <string-name>
            <surname>Auli</surname>
          </string-name>
          .
          <article-title>"Controllable abstractive summarization</article-title>
          .
          <source>" arXiv preprint arXiv:1711.05217</source>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Lloret</surname>
            , Elena, María Teresa Romá- Ferri, and
            <given-names>Manuel</given-names>
          </string-name>
          <string-name>
            <surname>Palomar</surname>
          </string-name>
          .
          <article-title>"COMPENDIUM: A text summarization system for generating abstracts of research papers." Data &amp; Knowledge Engineering (</article-title>
          <year>2013</year>
          ):
          <volume>88</volume>
          , pp.
          <fpage>164</fpage>
          -
          <lpage>175</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Ferrández</surname>
            , Oscar, Daniel Micol, Rafael Munoz, and
            <given-names>Manuel</given-names>
          </string-name>
          <string-name>
            <surname>Palomar</surname>
          </string-name>
          .
          <article-title>"A perspective-based approach for solving textual entailment recognition."</article-title>
          <source>In Proceedings of the ACL-PASCAL workshop on textual entailment and paraphrasing</source>
          , (
          <year>2007</year>
          ). pp.
          <fpage>66</fpage>
          -
          <lpage>71</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Gardenfors</surname>
            ,
            <given-names>Peter.</given-names>
          </string-name>
          <article-title>The geometry of meaning: Semantics based on conceptual spaces</article-title>
          . MIT press,
          <year>2014</year>
          . [10]
          <string-name>
            <surname>Luhn</surname>
            ,
            <given-names>Hans</given-names>
          </string-name>
          <string-name>
            <surname>Peter</surname>
          </string-name>
          .
          <article-title>"The automatic creation of literature abstracts." IBM Journal of research and development (</article-title>
          <year>1958</year>
          ): 2, pp.
          <fpage>159</fpage>
          -
          <lpage>165</lpage>
          . [11]
          <string-name>
            <surname>Saggion</surname>
            ,
            <given-names>Horacio.</given-names>
          </string-name>
          <article-title>"Learning predicate insertion rules for document abstracting."</article-title>
          <source>In International Conference on Intelligent Text Processing and Computational Linguistics</source>
          , Springer (
          <year>2011</year>
          ) pp.
          <fpage>301</fpage>
          -
          <lpage>312</lpage>
          . [12]
          <string-name>
            <surname>Maynard</surname>
            , Diana, Valentin Tablan, Hamish Cunningham, Cristian Ursu, Horacio Saggion, Kalina Bontcheva, and
            <given-names>Yorick</given-names>
          </string-name>
          <string-name>
            <surname>Wilks</surname>
          </string-name>
          .
          <article-title>"Architectural elements of language engineering robustness." Natural Language Engineering (</article-title>
          <year>2002</year>
          ): 8, pp.
          <fpage>257</fpage>
          -
          <lpage>274</lpage>
          . [13]
          <string-name>
            <surname>Witten</surname>
            , Ian H., and
            <given-names>Eibe</given-names>
          </string-name>
          <string-name>
            <surname>Frank</surname>
          </string-name>
          .
          <article-title>"Data mining: practical machine learning tools and techniques with Java implementations." ACM Sigmod Record (</article-title>
          <year>2002</year>
          ):
          <volume>31</volume>
          , pp.
          <fpage>76</fpage>
          -
          <lpage>77</lpage>
          . [14]
          <string-name>
            <surname>Sethi</surname>
          </string-name>
          , Nandini, Prateek Agrawal, Vishu Madaan, and Sanjay Kumar Singh.
          <article-title>"A novel approach to paraphrase Hindi sentences using natural language processing"</article-title>
          <source>Indian Journal of Science and Technology</source>
          (
          <year>2016</year>
          ):
          <volume>9</volume>
          (
          <issue>28</issue>
          ), pp.
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          . [15]
          <string-name>
            <surname>Alguliyev</surname>
          </string-name>
          ,
          <string-name>
            <surname>Rasim</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ramiz</surname>
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Aliguliyev</surname>
          </string-name>
          ,
          <string-name>
            <surname>Nijat R. Isazade</surname>
          </string-name>
          , Asad Abdi, and Norisma Idris. “COSUM:
          <article-title>Text summarization based on clustering and optimization</article-title>
          .
          <source>” Expert Systems</source>
          (
          <year>2019</year>
          ):
          <volume>36</volume>
          , doi:10.1111/exsy.12340. [16]
          <string-name>
            <surname>Gupta</surname>
            , Monika, and
            <given-names>Parul</given-names>
          </string-name>
          <string-name>
            <surname>Gupta</surname>
          </string-name>
          .
          <article-title>"Research and implementation of event extraction from twitter using LDA and scoring function."</article-title>
          <source>International Journal of Information Technology</source>
          (
          <year>2019</year>
          ):
          <volume>11</volume>
          , pp.
          <fpage>365</fpage>
          -
          <lpage>371</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [17]
          <string-name>
            <surname>Nizami</surname>
            , Muhammad, and
            <given-names>Ayu</given-names>
          </string-name>
          <string-name>
            <surname>Purwarianti</surname>
          </string-name>
          .
          <article-title>"Modification of ChuLiu/Edmonds algorithm and MIRA learning algorithm for dependency parser on Indonesian language."</article-title>
          <source>In 2017 International Conference on Advanced Informatics, Concepts</source>
          , Theory, and
          <string-name>
            <surname>Applications</surname>
          </string-name>
          (ICAICTA),
          <source>IEEE</source>
          (
          <year>2017</year>
          ) pp.
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [18]
          <string-name>
            <surname>Malik</surname>
            , Sonika, and
            <given-names>Sarika</given-names>
          </string-name>
          <string-name>
            <surname>Jain</surname>
          </string-name>
          .
          <article-title>"Ontology based context aware model."</article-title>
          <source>In 2017 International Conference on Computational Intelligence in Data Science (ICCIDS)</source>
          ,
          <source>IEEE</source>
          (
          <year>2017</year>
          ). pp.
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [19]
          <string-name>
            <surname>Sanchez-Gomez</surname>
            ,
            <given-names>Jesus M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miguel</surname>
            <given-names>A</given-names>
          </string-name>
          .
          <string-name>
            <surname>Vega-Rodríguez</surname>
            , and
            <given-names>Carlos J.</given-names>
          </string-name>
          <string-name>
            <surname>Pérez</surname>
          </string-name>
          .
          <article-title>"Extractive multi-document text summarization using a multi-objective artificial bee colony optimization approach." Knowledge-Based Systems (</article-title>
          <year>2018</year>
          ):
          <volume>159</volume>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [20]
          <string-name>
            <surname>ShafieiBavani</surname>
            , Elaheh, Mohammad Ebrahimi, Raymond Wong, and
            <given-names>Fang</given-names>
          </string-name>
          <string-name>
            <surname>Chen</surname>
          </string-name>
          .
          <article-title>"A graph-theoretic summary evaluation for rouge."</article-title>
          <source>In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing</source>
          (
          <year>2018</year>
          ): pp.
          <fpage>762</fpage>
          -
          <lpage>767</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [21]
          <string-name>
            <surname>Anand Kumar</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Singh</surname>
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kavirajan</surname>
            <given-names>B.</given-names>
          </string-name>
          and
          <string-name>
            <surname>Soman</surname>
            <given-names>K.P.</given-names>
          </string-name>
          “
          <article-title>Shared Task on Detecting Paraphrases in Indian Languages (DPIL): An Overview”</article-title>
          . In: Majumder P.,
          <string-name>
            <surname>Mitra</surname>
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mehta</surname>
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sankhavara</surname>
            <given-names>J</given-names>
          </string-name>
          . (eds)
          <article-title>Text Processing</article-title>
          .
          <source>FIRE 2016. Lecture Notes in Computer Science</source>
          , Springer (
          <year>2018</year>
          ):
          <fpage>10478</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [22]
          <string-name>
            <surname>Alphonsa</surname>
            ,
            <given-names>MM</given-names>
          </string-name>
          <string-name>
            <surname>Annie</surname>
            , and
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Amudhavalli</surname>
          </string-name>
          .
          <article-title>"Genetically modified glowworm swarm optimization based privacy preservation in cloud computing for healthcare sector</article-title>
          .
          <source>" Evolutionary Intelligence</source>
          (
          <year>2018</year>
          ):11 pp:
          <fpage>101</fpage>
          -
          <lpage>116</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [23]
          <string-name>
            <surname>Janani</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , and
          <string-name>
            <given-names>S.</given-names>
            <surname>Vijayarani</surname>
          </string-name>
          .
          <article-title>"Text document clustering using spectral clustering algorithm with particle swarm optimization." Expert Systems with Applications</article-title>
          ,
          <source>Elsevier</source>
          (
          <year>2019</year>
          ) 134, pp.
          <fpage>192</fpage>
          -
          <lpage>200</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [24]
          <string-name>
            <surname>Xu</surname>
            , Song,
            <given-names>Haoran</given-names>
          </string-name>
          <string-name>
            <surname>Li</surname>
            ,
            <given-names>Peng</given-names>
          </string-name>
          <string-name>
            <surname>Yuan</surname>
            , Youzheng Wu, Xiaodong He, and
            <given-names>Bowen</given-names>
          </string-name>
          <string-name>
            <surname>Zhou</surname>
          </string-name>
          .
          <article-title>"Self-Attention Guided Copy Mechanism for Abstractive Summarization." In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics (</article-title>
          <year>2020</year>
          ): pp.
          <fpage>1355</fpage>
          -
          <lpage>1362</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [25]
          <string-name>
            <surname>Weng</surname>
          </string-name>
          ,
          <string-name>
            <surname>Shi-Yan</surname>
          </string-name>
          , Tien-Hong
          <string-name>
            <surname>Lo</surname>
          </string-name>
          , and Berlin Chen.
          <article-title>"An Effective Contextual Language Modeling Framework for Speech Summarization with Augmented Features."</article-title>
          <source>In 2020 28th European Signal Processing Conference (EUSIPCO)</source>
          ,
          <source>IEEE</source>
          (
          <year>2021</year>
          ): pp.
          <fpage>316</fpage>
          -
          <lpage>320</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [26]
          <string-name>
            <surname>Mallick</surname>
          </string-name>
          , Chirantana,
          <string-name>
            <surname>Ajit Kumar Das</surname>
          </string-name>
          ,
          <string-name>
            <surname>Madhurima Dutta</surname>
          </string-name>
          ,
          <string-name>
            <surname>Asit Kumar Das</surname>
            , and
            <given-names>Apurba</given-names>
          </string-name>
          <string-name>
            <surname>Sarkar</surname>
          </string-name>
          .
          <article-title>"Graph-based text summarization using modified TextRank." In Soft computing in data analytics</article-title>
          , Springer (
          <year>2019</year>
          ): pp.
          <fpage>137</fpage>
          -
          <lpage>146</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [27]
          <string-name>
            <surname>Sabbah</surname>
          </string-name>
          , Thabit, Ali Selamat, Md Hafiz Selamat, Fawaz S.
          <string-name>
            <surname>Al-Anzi</surname>
            , Enrique Herrera Viedma, Ondrej Krejcar, and
            <given-names>Hamido</given-names>
          </string-name>
          <string-name>
            <surname>Fujita</surname>
          </string-name>
          .
          <article-title>"Modified frequencybased term weighting schemes for text classification." Applied Soft Computing (</article-title>
          <year>2017</year>
          ):
          <volume>206</volume>
          , pp.
          <fpage>58</fpage>
          <lpage>193</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [28]
          <string-name>
            <surname>Tofighy</surname>
            ,
            <given-names>Seyyed</given-names>
          </string-name>
          <string-name>
            <surname>Mohsen</surname>
          </string-name>
          ,
          <article-title>Ram Gopal Raj, and Hamid Haj Seyyed Javad. "AHP techniques for Persian text summarization."</article-title>
          <source>Malaysian Journal of Computer Science</source>
          (
          <year>2013</year>
          ):
          <volume>26</volume>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [29]
          <string-name>
            <surname>Bansal</surname>
            , Neha,
            <given-names>Arun</given-names>
          </string-name>
          <string-name>
            <surname>Sharma</surname>
            , and
            <given-names>R. K.</given-names>
          </string-name>
          <string-name>
            <surname>Singh</surname>
          </string-name>
          .
          <article-title>"Fuzzy AHP approach for legal judgement summarization</article-title>
          .
          <source>" Journal of Management Analytics</source>
          (
          <year>2019</year>
          ):6, pp.
          <fpage>323</fpage>
          -
          <lpage>340</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [30]
          <string-name>
            <surname>Gupta</surname>
            , Charu,
            <given-names>Amita</given-names>
          </string-name>
          <string-name>
            <surname>Jain</surname>
            , and
            <given-names>Nisheeth</given-names>
          </string-name>
          <string-name>
            <surname>Joshi</surname>
          </string-name>
          .
          <article-title>"Fuzzy logic in natural language processing-a closer view." Procedia computer science (</article-title>
          <year>2018</year>
          ):132 pp.
          <fpage>1375</fpage>
          -
          <lpage>1384</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>