=Paper= {{Paper |id=Vol-2823/Paper17 |storemode=property |title=Context-Aware Model of Abstractive Text Summarization for Research Articles |pdfUrl=https://ceur-ws.org/Vol-2823/Paper17.pdf |volume=Vol-2823 |authors=Gopinath Dineshnath, Selvaraj Saraswathi }} ==Context-Aware Model of Abstractive Text Summarization for Research Articles== https://ceur-ws.org/Vol-2823/Paper17.pdf
Context-Aware Model of Abstractive Text Summarization for
Research Articles
Gopinath Dineshnatha, Selvaraj.Saraswathib
    a
      Department of Computer Science Engineering, Pondicherry Engineering College, Pillaichavady,
    Puducherry, 605014, India.
    b
      Department of Information Technology, Pondicherry Engineering College, Pillaichavady,
    Puducherry, 605014, India.


     Abstract
     Research article comprises of different sections each holds its own characteristic domain information.
     Summarization of entire article from multiple documents of multiple sections in precise form with
     special focus to contextual information is tedious. We proposed context-aware model to extract
     contextual texts from research article by utilizing multi-document directed graph for contextual
     matching phrases We customized extractive summarization for abstractive text summarization with
     lucid-information as prime criterion. Decision matrix with elitism identification further fine tunes the
     abstractive text summary and outperforms at sentence level Rouge-L measures 9.32 and summary level
     measures 89.65.
     Keywords1
     Glowworm Swarm Optimization, Analytical Hierarchical Processing (AHP), Skip Gram Model,
     Word2vectors, Maximum Marginal Relevance, Log-Likelihood Rate,N-Gram.

1. Introduction                                                                            in two ways; single-document summaries
                                                                                           produce a summary from a given single source
                                                                                           and multi-document summaries in which
Modern days, internet becomes the integral part
                                                                                           different but related documents are summarized
of human and act as information highway. The
                                                                                           by comprises only the vital materials or main
primary source of information in digital world
                                                                                           ideas in a document in less space.
is Internet and it is boon for academicians,
                                                                                            There is a vast difference between automatic
bloggers, students and researcher fraternity.
                                                                                           multi-document summarization of generic texts
Information available in Internet comprises of
                                                                                           to that of scientific articles. The major
massive flow of information, which makes
                                                                                           difference [2] between generic-text and research
retrieval process complex with respect to
                                                                                           article summarization is; research article
context-specific content. Scientific article
                                                                                           consists of different section namely abstract,
prevailing now with ocean of research domains
                                                                                           introduction, literature survey, methods, results
makes difficult to scholar cope-up, grasp and
                                                                                           and discussions, whereas generic text’s scope is
streamline documents relevant to their interest.
                                                                                           extracted from first few sentences in first
Query based search [1] for specific domain also
                                                                                           paragraphs and entire section holds at
fetch many relevant articles that is difficult task
                                                                                           maximum 500 words.
to categorize surpass human processing
                                                                                               In general, abstract and citation texts in
capabilities. In such scenario, automatic text
                                                                                           scientific articles are considered for automatic
summarization of articles is fruitful solution in
                                                                                           summarization system.
terms of reducing time effort for reviewing
                                                                                               Abstract section[3] is biased to author
entire articles and grab gist of information
                                                                                           findings, author’s own contribution, and
enclosed in it. Basically, summaries generation
                                                                                           evaluation metrics. In simpler way, abstract

 ACI’21: Workshop on Advances in Computational Intelligence
 at ISIC 2021, February 25-27,2021, Delhi, India.
 Email: dinesh.gopinath60@gmail.com(G.Dineshnath);
 swathi@pec.edu (S.Saraswathi);
         :0000-0003-0026-1932(A.1)
            ©️2021 Copyright for this paper by its authors. Use permitted under Creative
            Commons License Attribution 4.0 International (CC BY 4.0).

            CEUR Workshop Proceedings (CEUR-WS.org)
outlines the domain and list of findings in crisp    techniques. Our main focus is Context aware
manner depends upon the type of articles             information    inference   from    multiple
(review/original).                                   documents.
Citation sensitive or citation-based summaries
[4]
     is another type of Scientific article           2. Literature review
summarization, major task in summary
production is clear-cut distinction of cited and
                                                     Lloret et al. [7] have applied both extractive and
non-cited text is performed. Citation summary
                                                     abstractive summarization procedures for
system performs categorization of every
sentence and labeled to citation or non-citation     scientific article abstracts.      The extractive
one. Later, evaluation measures based on             summarizer (compendium E) is developed to
similarity between each sentence in the              perform conventional preprocessing such as
reference article and citation sentences and then    breaking       sentences,     assigning    tokens,
grouped it into one of two classes: cited or non-    stemming, lemmatization and PNG markers,
cited.      Abstractive      Multi      document     tagging and removing duplicates at various
                                                     sentence levels. [8]A mixture of both extractive
summarization [5] selectively picks up either
                                                     and abstractive technique (compendium E−A)
first sentence of abstract or introduction of a
                                                     is developed to support compendium E as base
paper since it is comprised of background
                                                     to incorporates sorted information which are
information of research topic. Construction of
appropriate title for article involves               relevant. Relevancy identification with respect
interpretation and integration of the concepts       to every sentences, assigns a score that
from multiple sentences of the abstract. Apart       emphasize its importance based on code
from that there exist multiple challenging issues    quantity principle (CQP)[9] compendium E−A
like      content     organization,      sentence    derives abstractive summary by utilizing top
compression and fusion, and paraphrasing             ranked sentences with chronological ordering.
                                                               Saggion [11] utilized pretrain models
sentences.
All summarization system should meet the             for learning and transformations for the
summary length constraints as and other              problem of abstract generation. The initial
parameters specified by the user or                  summary generation from abstracts are
                                                     generated and transformed to model based
summarization system is known as controllable
                                                     learning. The learning models assists with
summarization. Controllable summarization [6]
is the main criteria in summarization system         examples from corpus. Further, abstracts are
which specifies the length of summary                gathered from GATE [12]and Weka [13]
generation in accordance to the entities on          environment. Abstractive text summarization
which it focuses on and mimics the source’s          also known as natural language generation in
                                                     natural language processing Paraphrasing of
style. User may define the high-level attributes
for summary generation. Summary generation           sentences is also another important criterion in
is controlled by specific control variables,         natural language generation Paraphrasing of
length, and source style, entities of interest and   sentences [14] involves substitution of relevant
summarizing only remaining portions of the           verbatim and modifying the direct to indirect
document. For instance, blog summarization,          speech or vice versa. The vector representation
the primary thing is to derive representative        is purely focus on various sources of features
words from comments and then selection of            namely LDA, D2V, W2V and encoding
                                                     schemes[15]..LDA[16]        explores     semantic
paramount sentences from the blog post which
consists of representative words.                    associations, D2V vectors finds contextual
    Context aware components are usually             word vectors along with documents. Contextual
meant task of inferring contextual information.      aware model phase is concerned with
Contextual information detection might be            contextual theme and dependency phrase
detection ranging from topic community,              extraction from multi documents using directed
paragraphs analysis, sentences and words by          graph. minimal spanning tree is constructed for
statistical computation measures. The most           edges algorithm using Chu Liu Edmonds.
well-known computation is Set-Calculus               (CLE)[17] Knowledge base or ontology- based
techniques to find sentence similarity via           context aware model useful for domain
Union, Jaccard Co-efficient, Cosine -Similarity      classification and annotation.[18]Evaluation of
                                                     document           summarization,       Document
measures       followed      by     normalization
Understanding        conference     (DUC)[19]                                      is used to slide over word vectors for prediction
benchmark datasets is used generally. Various                                      of context vectors. Skip gram models skips the
datasets such as TAC[20],shared task for text                                      selection of common words rather than
processing and document summarization                                              contextual words. N gram models hold
Similarly, DPIL[21] for paraphrasing Indian                                        predefined size ‘n’ that triggers the selection of
languages for text summarization.                                                  contextual words upon size limited to ‘n’. Both
                                                                                   skip gram and N gram models are desirable
                                                                                   notion to pick keywords in context based on
                    W(Vi) = (1 – d) + d ∗                                          their lexicography collocations. The Proposed
                              overlap(si ,sj )                             (1)
           ∑Vj∈ln(Vi)                                W(Vj)                         diagram is given in figure-1.
                        ∑ Vk∈Out(Vj )overlap(sj ,sk)
     .

   The edge weights for nodes are assigned
using status score on the basis of inwards and
outwards edges, W(Vi) represents status score
assigned to vertex Vi. In (Vi) and Out (Vj) are
inwards and outwards edges, points from
particular node. After several iterations, each
sentence in the document is assigned with a
score. The top-n sentences are selected and
ranked which constructs the summary for the
document. There exist some dependent phrases
in graph. The dependent phrases are cyclic in
nature and some disjoint nodes. Such, disjoint
nodes in graph are connected using CLE
algorithm. On basis of lexical contexts,
proposed procedure performs well than
traditional keyword-based algorithms. We
enhance the extractive summary production by
adding co-occurrence measures to ensure
Concept-based Point wise Mutual Information.
(CPMI) CPMI weighs the different section in
paragraphs gradually it weights decreases from
beginning of paragraph to end of paragraph in
document.       CPMI      measures      support
distributional semantics among phrases. CPMI
weightage scheme H(ƥ) is expressed in
equation 2.

 H(ƥ)
                                                        log B                (2)
         C − positive constant ƥ ∗ B,         if ƥ < − (       )
                                                         log C
 =                                                                     ,

     {                      1     otherwise                        }
                                                                                   Figure 1: Proposed Context aware model for
                                                                                   Abstractive Summarization
3. Context Aware Model Phase
Context aware component model proposed                                             3.1.    Contextual matching phrases
follows contextual keyword interpretation,
topic detection and topic clusters formation.
                                                                                   Contextual matching phrase from multi
Context aware component model also
                                                                                   documents is proposed to retrieve thematic
determines vector manipulation by bag of
                                                                                   portions with similar sentence phrases
words and skip gram key terms with respect to
                                                                                   sequentially from one document to another. The
specific documents. One hot encoding scheme
                                                                                   graph based contextual word is intuitive way to
         Formula for Extraction of           Eq      customized to produce extractive to abstractive
                Features.                    n.o:    summary.in Table-1. The optimal features and
                Number of title Features     (3)     their convergence in Adaptive Glowworm
         𝑇𝑖 =
                  average length(Title)              Optimization (AGWO) is discussed in section
                           1 − e –α          (4)
                Tag sum =                            3.4.
                            1 + eα
                           ∑L(Si)            (5)
                   (Sbi) =
                           Lmax                         Table 1 Optimal Features Extraction.
                            TW(𝑆𝑖)           (6)
                Dsm(𝑆𝑖) =
                             𝑇𝑀𝑎𝑥
represent sentences as node and corresponding
contextual word as vertices. The path projecting                  Features            Glowworm
from node to node via outcoming and incoming                                         optimization
vertices provides a notion either matching                  Title feature          Luciferin update
target phrases or discriminating phrases.                                               phase
subsequent dependent phrases need to be                   Named entity             Movement phase
included in directed graph. The procedure for            recognition and
contextual theme and dependency phrases                      Tagging
extraction is shown below.                             Sentence boundary        Neighborhood Phase
   Step 1 Accept text and store it in text buffer.        Distributional        Words and conceptual
   Step 2 new word falls into below categories.
                                                        Semantic Analysis              level
   Step 2 a): If first word then adds to graph G.
   Step 2 b): If fresh word then appends to G.
   Step 3 Go to step 2 until words overlap.          3.3.     Vector Formulation
   Step 4 If overlap, status score using (1).
   Step 5 Extract similar texts and update in G.     Word embeddings are feature vectors that
   Step 6 Construct Digraph using CLE.               represent words holds the property that similar
   Step 7 Do updates to infer adjacent edges.        words have similar feature vectors. The
   Step 8: Output the phrases.                       question might rise in mind where the
                                                     embeddings come from. The response is
3.2.    Feature Extraction                           (again): they are learned from data. Several
                                                     algorithms exist to learn word embeddings. We
The feature extraction for document                  consider only one of them: word2vec, and sole
summarization includes title feature, proper         version of word2vec called skip-gram, which is
nouns, Named Entities Recognition (NER) and          well-known and currently utilized in practice.
parts of speech tagging, sentence boundary           Word embedding learning, our goal is to build
analysis and distributional semantic analysis.       a model which we can use to convert a one-hot
Title feature scoring scheme is based on the         encoding of a word into a word embedding. Let
ratio of mean number of titles present to that of    our dictionary contain 10,000 words or Giga-
average length of title. The formula for title       word Corpus. Skip-gram model performs for
feature is expressed in equation. (3). The proper    given sentence, selection of a word is feed into
nouns are generally recognized as title words        classifier, and predict words before and after the
and minimum number of words to accept as             selected word in a fixed window. Negative
title. NER marks or labels the salient sentences     sampling provides better vectors for frequent
which is considered for summary. The scoring         words with low dimension.
scheme is expressed in equation. (4). where α =          Latent Dirichlet (LD) allocation is a
(t(s) − µ) /σ (sigmoid function) aggregates          possibility-based mechanism viable for
mean count of regular expressions, case level        assortments, for example, text assortments. LD
and numeric literals. Sentence boundary              consolidates the documents as a blend of shifted
calculation is expressed in equation (5).            topics; every unit involves words that have a
Distributional Semantic Analysis weighs              spun affiliation that exists between them. Also,
thematic concepts to find word co-occurrence.        word choice simply dependent on the numerical
It is scored using formula. (6). Column one          idea of likelihood. Recursively determining the
represents features, similarly column two            interaction of themes and words is done for the
represents formulae used for computation and         phase of a lonely record or a large number of
column three indicates optimal features
documents. At long last, yields the record                 Glowworm         Optimization         (GO)[22]
which comprises of different subjects.                 comprises of three phases namely Luciferin
    LD allocation algorithm performs the               update phase, Neighborhood phase and
following:                                             movement phase. Adaptive Glowworm
    1) Determine the number N of words in the          Optimization (AGO) is proposed for tailor-
document concerning probability distribution is        made features to acquire vectors or extract
Poisson.                                               features to frame the summary. The
    2) Pick a merge of focuses for the report          optimization principle is based on five features
from a predefined set of K subjects as                 and their application phase is listed in Table-1.
demonstrated by the Dirichlet movement.                    Sentence position is additional feature to
    3) Produce individually word in the list of        revisit sentence with appropriate ordering.
terminology as follows language Vocabulary             Positioning of sentences which is most vital
(V).                                                   part in the summary generation have higher
         a) Choose a subject;                          weights. The feature associated with sentence
         b) Choose a word in this subject.             length; hence we have minimal set of 25 words
                                                       to accept as a sentence. F5 feature is in
3.4.     Sentence Ranking                              movement phase of Glow worm optimization
                                                       with lucerifin value or luminous quotient, affine
                                                       towards the similar topics.
Sentence ranking phase, chiefly performs                   Luciferin update phase, sentences are
identification of prominent sentences with             concatenated with respect to the relevancy.
pertinent information and free from                    Relevancy is determined by feature with
redundancy. It selects topmost sentences from          respect to title and all sentences in document.
documents and produces summary with                    Luciferin update phase, movement phase and
application of traditional maximization                neighbor phase are expressed as equation in (7)
algorithm like EM.[23]. finally produces
                                                       (8)    and      (9)    respectively     Luciferin
extractive summary. However, Extractive                enhancement(ʋ) depends upon Proper Nouns.
summary lacks positional placement of
sentences. hence there is a need to revisit               J(xi(t))-objective function which maximize
sentence positions.[24].
                                                       weights of every proper nouns. Luciferin decay
                                                       constant gradually decreases when common
3.5.     Extractive Summary                            noun exists.
                                                          Movement Phase, forms local clusters based
Extractive summary is created with top n               on decision range. Sentences are of similar
sentences for research articles summarization          contexts likely to move based on entity-tagging
Latter summary is transformed to decision              features. Finally, neighbor Phase performs
matrix. Decision Matrix will keep track of             chronological sorting of clustered sentences to
extractive summary to make compatible for              produces summary.
abstractive text summary with usage measure
and penalty measure. Hence, optimization               4. Results and Discussions
algorithm is used to remove exact replica of
original text produced in extractive summary.          The extractive summarization shows better
Co-reference resolution is also handled in             results with proposed procedure, the summary
abstractive    text    summary      generation.        produced purely relies on lexical features and
Meanwhile construction of cosine similarity            surpass traditional keyword ranking schemes.
was rapid, useful, and seems reasonable.               yellow color denotes dependency phrases and
                                                       green color denotes contextual theme. The
3.6.     Optimization                                  output of context aware component of
                                                       extractive summarization is shown as well as
                                                       output of abstractive text summarization is also
   li(t + 1) = (1 − 𝜌)li(t) + ʋJ (xi (t + 1));   (7)   shown.
    Ni(t) = {j: dij < rid(t); li (t) < lj (t)}   (8)

       𝑟𝑖𝑑(𝑡 + 1) = 𝑚𝑖𝑛{𝑟𝑠, 𝑚𝑎𝑥{0,               (9)
                          𝑟𝑖𝑑(𝑡) + 𝛽(ǹ𝑡
                          − |𝑁𝑝(𝑡)|}};
  4.1.     Training and Testing
                                                           Output of Context aware
  For training purpose, Document understanding             Extractive and Abstractive
  Conference (DUC) data set taken into                        Summarization
  considerations. The precise explanation of
  DUC data sets and DUC data is customized,
  which is free from least significant words or                Extractive Summary
  stop words according to port-stemmer’s               A large number of methods applied in
  algorithm. Recall oriented understudy Gist        the field of extractive summarization
  evaluation, Rouge(R) is also considered for       over the years. Scoring sentences for
  evaluation. R falls into many variants like R-    such summary is tedious task. Many
  unigram, R-bigram, R-Longest common
                                                    researchers putting so much effort to
  Subsequence and R-N gram classes.
  Multiple documents of artificial intelligence
                                                    improve the quality of summary.
  domain for testing and performed various          Document summarization focus both
  measures like R-1, R-2, and R-L (Longest          quality and coverage of content.
  Common Subsequence) scoring. At Sentence-         Clustering of sentences in document
  level, computes longest common subsequence        summarization shown promising results
  (LCS) between two pieces of text ignores new      to discover topics. A Fuzzy oriented
  lines and summary-level, newlines in the text     clustering for summarization of multi-
  are interpreted as sentence boundaries, and the   documents. Compendium- a summarizer
  LCS is computed between each pair of              tool, generates relevant summary free
  reference and candidate sentences, and their      from redundancy Collabsum clustering
  results are tabulated below in table-2.
                                                    process both inter and intra document
  The proposed AGO performs well in sentence
  and summary level than traditional methods.
                                                    relationship and forms clusters. Clusters
  similarly, contextual theme detection also out    in turn apply graph based ranking
  performs than traditional schemes like lexrank,   methodology.            FEOM…genetic
  maximum relevance, loglikelihood ranking          algorithm…graph                   -based
  [25],[26],[27]
                 and other centrality measures as   approach…probabilistic model.
  baseline evaluation.

             Table 2 Results Comparison                       Abstractive summary

Methods      R-1       R-2    R-L      R-L             Various types of Sentence clustering
                         Sentence Summary           techniques applied to document
                           level    level           summarization Sentence scoring, topic
Proposed   0.4611 0.1342   0.932    0.89            coverage, relevant sentences and
  AGO                                               summarization quality are main
  MMR      0.3716 0.0757         0.16        0.80   components in summary production.
   LEX     0.4030 0.0913         0.69        0.53   Clustering algorithms for sentence
                                                    scoring and grouping similar sentences
 Proposed 0.4297 0.0835          0.12        0.83   according to topics conveyed in
contextual                                          document. A fuzzy based, evolutionary
  theme                                             based clustering also successfully
   Tf-idf  0.3639 0.0736         0.14        0.81   applied in conjunction with other graph-
  LLR      0.3975 0.0850        0.084        0.64   based approaches to provide summary.
5. Conclusion                                      References
Abstractive text summarization for research             [1] Shafiei        Bavani,         Elaheh,
articles generates sentences individually using             Mohammad Ebrahimi, Raymond
glowworm optimization with six associated                   Wong, and Fang Chen. "A query-
features. In addition, decision matrix with                 based summarization service from
elitism identification is formulated to choose              multiple news sources." In 2016
summary sentences from both extractive                      IEEE International Conference on
summary sentences and abstractive summary                   Services Computing (SCC), IEEE,
sentences with consistency as necessary                     (2016):         pp.42-49.          doi:
condition. Extractive summary is reduced to                 10.1109/SCC.2016.13.
more than 80% to generate abstractive                   [2] Bharti, Santosh Kumar, Korra
summary. Extractive summary with 661 word                   Sathya Babu, Anima Pradhan, S.
tokens is produced as output in first phase.                Devi, T. E. Priya, E. Orhorhoro, O.
Later, decision matrix with Elitism                         Orhorhoro, V. Atumah, E. Baruah,
identification produces abstractive summary                 and P. Konwar. "Automatic
with 84 tokens is obtained as final output.                 keyword extraction for text
Proposed multi-document directed graph                      summarization in multi-document
contextual matching phrases, Rouge-L                        e-newspapers articles." European
measures in sentence level is 12.08 and Rouge-              Journal      of      Advances        in
L measures in summary level is 83.95 for                    Engineering and Technology
extractive summary. Similarly, Rouge-L                      (2017): 4, pp.410-427.
measures in sentence level is 9.32 and Rouge-L          [3] Chowdhury, S. M., and Mazharul
measures in summary level is 89.68 for                      Hoque. "A Review Paper on
abstractive summary. A novel model has been                 Comparison          of       Different
implemented to be ample enough to provide                   Algorithm        Used      in     Text
multi objectives and to convince the                        Summarization." Intelligent Data
instantaneous needs. Ultimately, this study will            Communication Technologies and
inspire many researchers to further explore and             Internet of Things: ICICI 2019 38
apply the various types of Swarm intelligence               (2019): 114.
while solving the summarization tasks,                  [4] Cohan,      Arman,       and     Nazli
specifically    in    the    abstractive    text            Goharian.       "Scientific     article
summarization (ATS) field.                                  summarization using citation-
                                                            context and article's discourse
6. Future works                                             structure." (2017). arXiv preprint
                                                            arXiv:1704.06619
                                                        [5] Nouf Ibrahim Altmami, Mohamed
    Decision matrix performs combination of                 El Bachir Menai, “Automatic
Sentences from extractive summary are                       summarization        of      scientific
assessed and deemed to be fit for abstractive               articles: A survey.” Journal of King
summary are analyzed in conjunction with                    Saud University-Computer and
input from optimization algorithm with                      Information Sciences, (2020)
associated six features. Selection of best                  doi.:10.1016/j.jksuci.2020.04.020.
sentences and worst sentences based on their            [6] Fan, Angela, David Grangier, and
usage and penalty is awarded to compose                     Michael       Auli.     "Controllable
summary. Global decision matrix performs                    abstractive summarization." arXiv
elitism identification (algorithm) and outputs              preprint arXiv:1711.05217 (2017).
sentences with sentence flow as criterion.              [7] Lloret, Elena, María Teresa Romá-
However, decision matrix follows Analytical                 Ferri, and Manuel Palomar.
Hierarchical Processing (AHP) [28] with user                "COMPENDIUM:               A       text
defined decision values and their decisions are             summarization         system        for
normalized. We can extend the normalized                    generating abstracts of research
vectors by using fuzzy [29] based membership                papers." Data & Knowledge
assessment as stated by Charugupta,et.al.[30]
          Engineering (2013):88, pp.164-         [17] Nizami, Muhammad, and Ayu
          175.                                       Purwarianti. "Modification of Chu-
[8] Ferrández, Oscar, Daniel Micol, Rafael           Liu/Edmonds algorithm and MIRA
    Munoz, and Manuel Palomar. "A                    learning algorithm for dependency
    perspective-based approach for solving           parser on Indonesian language." In
    textual entailment recognition." In              2017 International Conference on
    Proceedings of the ACL-PASCAL                    Advanced Informatics, Concepts,
    workshop on textual entailment and               Theory, and Applications (ICAICTA),
    paraphrasing, (2007). pp.66-71.                  IEEE (2017) pp.1-6.
  [9] Gardenfors, Peter. The geometry of         [18]Malik, Sonika, and Sarika Jain.
      meaning:      Semantics     based    on        "Ontology based context aware
      conceptual spaces. MIT press, 2014.            model."     In    2017    International
  [10]Luhn, Hans Peter. "The automatic               Conference       on      Computational
      creation of literature abstracts." IBM         Intelligence in Data Science (ICCIDS),
      Journal of research and development            IEEE (2017). pp.1-6.
      (1958): 2, pp.159-165.                     [19]Sanchez-Gomez, Jesus M., Miguel A.
  [11]Saggion, Horacio. "Learning predicate          Vega-Rodríguez, and Carlos J. Pérez.
      insertion     rules    for    document         "Extractive     multi-document      text
      abstracting."       In     International       summarization using a multi-objective
      Conference on Intelligent Text                 artificial bee colony optimization
      Processing       and     Computational         approach." Knowledge-Based Systems
      Linguistics, Springer (2011) pp.301-           (2018): 159, pp.1-8.
      312.                                       [20]ShafieiBavani, Elaheh, Mohammad
  [12]Maynard, Diana, Valentin Tablan,               Ebrahimi, Raymond Wong, and Fang
      Hamish Cunningham, Cristian Ursu,              Chen. "A graph-theoretic summary
      Horacio Saggion, Kalina Bontcheva,             evaluation for rouge." In Proceedings
      and Yorick Wilks. "Architectural               of the 2018 Conference on Empirical
      elements of language engineering               Methods in Natural Language
      robustness."      Natural     Language         Processing (2018): pp.762-767.
      Engineering (2002): 8, pp.257-274.         [21]Anand Kumar M., Singh S., Kavirajan
  [13]Witten, Ian H., and Eibe Frank. "Data          B. and Soman K.P. “Shared Task on
      mining: practical machine learning             Detecting Paraphrases in Indian
      tools and techniques with Java                 Languages (DPIL): An Overview”. In:
      implementations." ACM Sigmod                   Majumder P., Mitra M., Mehta P.,
      Record (2002):31, pp.76-77.                    Sankhavara J. (eds) Text Processing.
  [14]Sethi, Nandini, Prateek Agrawal, Vishu         FIRE 2016. Lecture Notes in Computer
      Madaan, and Sanjay Kumar Singh. "A             Science, Springer (2018):10478.
      novel approach to paraphrase Hindi         [22]Alphonsa, MM Annie, and P.
      sentences using natural language               Amudhavalli. "Genetically modified
      processing" Indian Journal of Science          glowworm swarm optimization based
      and Technology (2016):9(28), pp.1-6.           privacy     preservation    in   cloud
  [15]Alguliyev, Rasim M., Ramiz M.                  computing for healthcare sector."
      Aliguliyev, Nijat R. Isazade, Asad             Evolutionary Intelligence (2018):11
      Abdi, and Norisma Idris. “COSUM:               pp: 101-116.
      Text     summarization      based    on    [23]Janani, R., and S. Vijayarani. "Text
      clustering and optimization.” Expert           document clustering using spectral
      Systems                      (2019):36,        clustering algorithm with particle
      doi:10.1111/exsy.12340.                        swarm optimization." Expert Systems
  [16]Gupta, Monika, and Parul Gupta.                with Applications, Elsevier (2019) 134,
      "Research and implementation of event          pp.192-200.
      extraction from twitter using LDA and      [24] Xu, Song, Haoran Li, Peng Yuan,
      scoring function." International Journal       Youzheng Wu, Xiaodong He, and
      of Information Technology (2019):11,           Bowen Zhou. "Self-Attention Guided
      pp.365-371.                                    Copy Mechanism for Abstractive
                                                     Summarization." In Proceedings of the
        58th Annual Meeting of the
        Association      for     Computational
        Linguistics (2020): pp.1355-1362.
[25] Weng, Shi-Yan, Tien-Hong Lo, and
      Berlin Chen. "An Effective Contextual
      Language Modeling Framework for
      Speech Summarization with Augmented
      Features." In 2020 28th European Signal
      Processing Conference (EUSIPCO),
      IEEE (2021): pp.316-320.
[26] Mallick, Chirantana, Ajit Kumar Das,
      Madhurima Dutta, Asit Kumar Das, and
      Apurba Sarkar. "Graph-based text
      summarization        using       modified
      TextRank." In Soft computing in data
      analytics, Springer (2019): pp.137-146.
[27] Sabbah, Thabit, Ali Selamat, Md Hafiz
      Selamat, Fawaz S. Al-Anzi, Enrique
      Herrera Viedma, Ondrej Krejcar, and
      Hamido Fujita. "Modified frequency-
      based term weighting schemes for text
      classification." Applied Soft Computing
      (2017):206, pp.58 193.
[28] Tofighy, Seyyed Mohsen, Ram Gopal
      Raj, and Hamid Haj Seyyed Javad. "AHP
      techniques       for     Persian     text
      summarization." Malaysian Journal of
      Computer Science (2013): 26, pp. 1-8.
 [29] Bansal, Neha, Arun Sharma, and R. K.
      Singh. "Fuzzy AHP approach for legal
      judgement summarization." Journal of
      Management         Analytics    (2019):6,
      pp.323-340.
 [30] Gupta, Charu, Amita Jain, and Nisheeth
      Joshi. "Fuzzy logic in natural language
      processing–a closer view." Procedia
      computer science (2018):132 pp.1375-
      1384.