<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>A select and rewrite approach to the generation of related work reports</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ahmed AbuRa'ed</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Horacio Saggion</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>LaSTUS/TALN Group</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Universitat Pompeu Fabra</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Spain</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>name.surname}@upf.edu</string-name>
        </contrib>
      </contrib-group>
      <pub-date>
        <year>2021</year>
      </pub-date>
      <fpage>53</fpage>
      <lpage>68</lpage>
      <abstract>
        <p>A related work report is a text which integrates key information from a list of related scientific papers providing context to the work being presented. In this paper we study the automatic generation of related work reports using extractive and abstractive text summarization approaches. Our extractive approach scores the sentences of the scientific papers based on their citations, selecting top scored sentences from each scientific paper to be mentioned in the related work report. The sentences are then organized in the report according to the topic they belong to. In additional experiments we use top scored sentences from our extractive methods and rephrase them using pre-trained abstractive models that generate citation sentences. We discuss automatic and manual evaluation of the generated related work reports showing the viability of the proposed approaches.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Scientific Summarization</kwd>
        <kwd>Information Extraction from Scientific Literature</kwd>
        <kwd>Document Abstracting</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>For argument labeling in discourse parsing on the PDTB corpus, the related work can be classified
into two categories: locating parts of arguments and labeling full argument spans.
As a representative on locating parts of arguments, Wellner and Pustejovsky (2007) proposed
several machine learning approaches to identify the head words of the two arguments for discourse
connectives. Following this work, Elwell and Baldridge (2008) combined general and connective
specific rankers to improve the performance of labeling the head words of the two arguments.
Prasad et al. (2010) proposed a set of heuristics to locate the position of the Arg1 sentences for
inters entence cases.
….….….</p>
      <p>In comparison, labeling full argument spans can provide a complete solution to argument labeling
in discourse parsing and has thus attracted increasing attention recently, adopting either a
subtree extraction approach (Dinesh et al. (2005), Lin et al. (2014)) or a linear tagging approach
(Ghosh et al. (2011)).</p>
      <p>As a representative subtree extraction approach, Dinesh et al. (2005) proposed an automatic tree
subtraction algorithm to locate argument spans for intra-sentential subordinating connectives.
….….….</p>
      <p>Instead, Lin et al. (2014) proposed a two-step approach. First, anargument position identifier
was employed ….
….….….</p>
      <p>As a representative linear tagging approach, Ghosh et al. (2011) cast argument labeling as a
linear tagging task using conditional random fields. Ghosh et al. (2012) further improved the
performance with integration of the n-best results.</p>
      <p>
        According to previous research [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ], related work reports are classified into descriptive
or integrative: a descriptive report will summarize individual papers providing information
such as methods and results in citation sentences. Instead, integrative reports will focus
on key ideas and topics, providing in the citation sentences critical views on the presented
approaches.
      </p>
      <p>In this paper we are concerned with the automatic production of descriptive related
work sections from a set of selected papers by using extractive and abstractive
(sequenceto-sequence) methods. The extractive method to be presented, which achieved state of
the art performance in citation-based summarization, learns to score sentences using a
Convolutional Neural Network (CNN). We use the top ranked sentences of each paper to
generate our extractive report. In order to account for coherence phenomena observed
in this type of text (See Figure 1), we propose a topic-based modelling approach for
information ordering. Moreover, in order to generate sentences matching the related work
report style we rely on an abstractive method based on Bidirectional Recurrent Neural
Networks (BRNN) and trained to generate citation sentences. The sentences extracted
with the CNN methods are feed into the BRNN to produce an abstractive related work
report. As it will be shown, the combination of extractive and abstractive techniques
improves the results in terms of automatic evaluation.</p>
      <p>The rest of the paper is organized as follows: in Section 2 we report related work in the
area, then in Section 3 we explain the methodology and data used. Section 4 presents the
experiments and Section 5 discusses automatic and human evaluation results. Finally,
Section 6 closes the paper.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        In contrast with generic summarization, state of the art generation and summarization
has not been extensively explored. Key works in the area are: [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] and [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], making [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]
to be the first to generate related work sections from a hierarchical topic-biased tree,
and [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] who deal with multi-document scientific article summarization. Other studies
investigate mainly single document scientific article summarization. In respect to that,
we will cover two main types of related works: Automated related work summarization
and Automatic text summarization (ATS) in the domain of scientific texts.
      </p>
      <sec id="sec-2-1">
        <title>2.1. Automatic text summarization (ATS) in the domain of scientific texts</title>
        <p>
          Although research in summarization can be traced back to the 50s [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] and even though
a number of important discoveries have been produced in this area, automatic text
summarization still faces many challenges given its inherent complexity. Scientific
text summarization is of paramount importance and scientific texts were automatic
summarization’s first application domain [
          <xref ref-type="bibr" rid="ref6 ref7">6, 7</xref>
          ]. Several methods and techniques have
already been reported in the literature to produce text summaries by automatic means
[
          <xref ref-type="bibr" rid="ref8">8</xref>
          ].[
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] tackled the multi-document summarization of scientific articles problem by an
original unsupervised method, in which the source document cites a list of papers (also
known as a co-citation). From each co-cited article, a topic based clustering of fragments
was mined and ranked using a query produced from the context surrounding the co-cited
list of papers. [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] proposed a model which uses a clustering approach to summarize a
single topic from the article and this summarized topic is further used to summarize the
entire topic of the specified article. The main contribution is to use citation summaries
and network analysis techniques which yield a summary of a single scientific article as a
framework for future research on topic summarization. [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] proposed a summarization
approach for scientific articles which takes advantage of citation-context and the document
discourse model. They also leverage the inherent scientific article’s discourse for producing
better summaries. [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] suggested that performing reinforcement ranking on the Semantic
Link Network of various representation units within a scientific paper (word, sentence,
paragraph and section) can significantly improve extractive summarization of paper. [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ]
proposed an approach to generate automatic summarization based on 5W1H (who, what,
whom, when, where, how) event structure. Sentences in literature are classified and
selected for diferent elements of events by relevance, and then, the importance of each
candidate sentence is calculated. Top-k relevant and important sentences are selected to
formulate event-based summarization.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Automated related work summarization</title>
        <p>
          [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] and [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ] presented the novel problem of automatic related work summarization. A
related work summarization system creates a topic-biased summary of related work for a
target paper given multiple scientific articles together with a topic hierarchy tree as an
input. [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] also stated that three things should be considered to generate a summary. First,
a mandatory input is needed for the summarization process identified as a high-level
rhetorical structure in a form of a topic tree. Second, summaries can be seen as transitions
along the topic hierarchy tree. Third, sentences either describe generic or specific topics.
Generic topics are often characterized by background information. This include definitions
or descriptions of a topic’s purpose. In contrast, detailed information forms the substance
of the summary and often describes key related work that is attributable to specific
authors.
        </p>
        <p>
          [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] investigated on the task of producing a related work section for a target paper,
provided a set of Reference Papers along with a target academic paper which has no
related work section as input. They developed an Automatic Related Work Generation
system (ARWG) that exploits the Probabilistic Latent Semantic Analysis (PLSA) [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ] to
solve this problem. They used the PLSA model to divide the sentence set of the given
papers into diferent topic-biased parts, and then applies regression models to learn the
standing (ranking) of the sentences. Finally, it utilizes an optimization framework to
produce the related work section. Their evaluation results on a test set of 150 target
papers sideways with their Reference Papers show that ARWG can indeed generate
related work sections with improved quality than those of baseline methods MEAD [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ]
and LexRank [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. A user study is also carried on to demonstrate that ARWG can
achieve improvements over generic multi-document summarization baselines. It is worth
noting that in this work they use abstract, introduction, related work and conclusion
sections, since other sections corresponding to method and evaluation sections always
describe in too much details the specific work.
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Method</title>
      <p>
        We score the sentences of the scientific papers based on their citation network selecting
those which score higher, then we generate an organized related work report based on the
relation between the scientific papers being summarized. In order to score the sentences
we use both supervised and unsupervised approaches. For the unsupervised learning we
use two methods: one is based on a modified variant of Jaccard Similarity, and the other
is based on the BabelNet Embeddings Distance. As for the supervised approach we use
several variations of a Convolutional Neural Network (CNN) [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. Once we have scored
the sentences, we select a unified number of sentences from each scientific paper and add
them to the final related work report in topic order. We use Latent Dirichlet Allocation
(LDA) [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] to perform topic modeling across the scientific papers detecting any cross
document linking between the scientific papers based on their topics.
      </p>
      <p>
        For additional experiments we rephrase the selected sentences using pretrained
sequenceto-sequence models trained with citation-sentences to paraphrase the extracted sentences
in a citation style.
3.1. Data
In order to generate related work reports through extractive summarization of scientific
papers we utilize a Multi-level Annotated Corpus of Scientific Papers. The corpus is
proposed by [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] which expands considerably the data-set of related work sections used
in [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] by providing: (i) related work sections, (ii) a manually annotated layer of cited
papers and sentences, (iii) citing papers referring to the cited papers in the related
work section, and (iv) a layer of rich linguistic, rhetorical, and semantic annotations
computed automatically. The corpus contains three types of scientific papers: target
papers, reference papers, and citing papers which are organized in a two-level network.
Level 1 contains target papers with their related work sections which cite a set of reference
papers. Level 2 extends the corpus by adding a layer representing a set of scientific
papers explicitly citing the reference papers in Level 1. This data-set is ideal for our
research: we use Level 2 of the corpus to score the reference papers’ sentences through
their connection with the scientific papers citing them in the citation network. Next, we
select the top sentences of the reference papers and generate an organized related work
report using Level 1 of the corpus, in which we try to re-create the related work section
of the target paper by summarizing the reference papers mentioned in it. Finally, for
evaluation we use the gold related work sections of the target papers provided by the
corpus.
      </p>
      <sec id="sec-3-1">
        <title>3.2. Scoring sentences of the Reference Papers</title>
        <p>Researchers tend to cite the major contributions of a scientific paper. Therefore, utilizing
the citation network between the scientific paper and the papers that are citing it will
provide an insight of what those researchers consider an important context in the scientific
paper.</p>
        <p>
          We performed experiments using two unsupervised methods using two similarity
metrics: Modified Jaccard similarity and BabelNet (a multilingual lexicalized semantic
network and ontology) Embeddings Distance [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ]. We used a metric similar to the
Jaccard similarity coeficient [
          <xref ref-type="bibr" rid="ref21">21</xref>
          ] for comparing two sentences (i.e., the citation sentence
of a citing paper with every sentence within a reference paper). This metric considers
the union and intersection of words (like the Jaccard coeficient) but uses the inverted
frequency information to give more weight to words in the intersection that are less
common. Our modification assigns greater weight to matching words that are infrequent
in the corpus, based on the idea that two text spans that share infrequent words are
more likely to be semantically related. The modified Jaccard similarity between two text
spans  1 and  2 is defined in Equation 1.
        </p>
        <p>(
1,  2) =
∑∈ 1∩ 2 2</p>
        <p>()
| 1 ∪  2|</p>
        <p>
          As for the second method we obtained the BabelNet synsets for both sentences and
transformed them into synset embeddings [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ]. We then take the cosine similarity between
the centroids of the synset embeddings for both the reference and citation sentences.
Once we have collected the similarities based on these two metrics between each citation
context in a citing paper and each sentence in the reference paper they cite, we formed
the final score for each sentence in the reference paper. The final score takes into account
the similarity of the citation context in the citing paper with the sentence in the reference
(1)
paper alongside the assigned weight of that citing paper. The weight for each scientific
paper is based on the number of scientific papers citing it.
        </p>
        <p>
          On the other hand, for our supervised approach, we needed a source of data to train
our CNN models. To do so we make use of a data set provided by the CL-Scisumm
Shared challenge [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ] which addresses the problem of summarizing a scientific paper
taking advantage of its citation network. The challenge organizers provided a cluster of
 documents where one is a reference paper (RP) and the  − 1 remaining documents
are papers (i.e., citing papers (CPs)) citing the reference paper, they also provide three
gold summaries for each reference paper alongside manual annotations stating which
sentences in the reference paper have been cited by the citation context of the citing
papers. The three types of summaries for each Reference Paper are:
• the abstract, written by the authors of the research paper.
• the community summary, collated from the majority of the reference spans of its
citances.
• a human-written summary, written by the annotators of the CL-SciSumm
annotation efort.
        </p>
        <p>
          The CNN scores the sentences of the scientific paper using linguistic and semantic
features from the paper itself alongside the papers that are citing it. The aim of our
CNN is to learn the relation between a sentence and a scoring value indicating its
relevance. Since the CL-Scisumm Challenge Dataset provides several types of gold
standard summaries, we use them to train several CNN models. The CNN learns the
relation between features and a score, that is regression [
          <xref ref-type="bibr" rid="ref24">24</xref>
          ]. The scoring functions are
defined below:
• Cosine Distance: we calculated the maximum cosine similarity between each
sentence vector in the Reference Paper with each vector in the gold standard
summaries. This method produced three scoring functions (based on SUMMA [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ]
word vectors, ACL embeddings, and Google embeddings) for each summary type.
• ROUGE-2 Similarity: we also calculated similarities based on the overlap of bigrams
between sentences in the Reference Paper and gold standard summaries. In this
regard, each sentence in the Reference Paper is compared with each gold standard
summary using ROUGE-2 [
          <xref ref-type="bibr" rid="ref26">26</xref>
          ].
• Scoring Functions Average: Moreover, we computed the average between all cosine
scoring functions (SUMMA, ACL, Google and ROUGE-2; referred to as SGAR
in the tables) for each summary type. In addition, we also calculated a simplified
average with vectors that are not based on word-frequencies (ACL, Google and
ROUGE-2; referred to as GAR in the tables).
        </p>
        <p>
          We have used the same set of features that [
          <xref ref-type="bibr" rid="ref27">27</xref>
          ] used at their participation in the
CLSciSumm challenge which achieved state of the art performance. After training the
CNN models, we pass the reference papers of the Multi-level Annotated Corpus (Section
3.1) as testing data to score their sentences. Once we have scored the sentences of the
reference papers using the supervised and unsupervised methods, we sort the sentences
in descending order before moving to selecting and organizing the final output.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.3. Generating the Related Work Report</title>
        <p>
          Since authors of related work reports usually starts with a certain related topic and
move onward stating each and every reference paper related to that specific topic before
moving to the next topic, we group the sentences of reference papers that share the
same topic together. To find the topics across the reference papers we used Latent
Dirichlet Allocation (LDA) [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ] and modeled each reference paper based on its Title and
Abstract. In order to find the optimal number of topics to train the LDA model on, we
build many LDA models based on diferent number of topics (  ) and pick the one
that gives the highest coherence value or till the coherence value converges, choosing a
‘ ’ that marks the end of a rapid growth of topic coherence usually ofers meaningful
and interpretive topics, while picking an even higher value can sometimes provide more
granular sub-topics.
        </p>
        <p>Once we identify the LDA model with the ideal number of topics to train on ( ),
we use it to identify the topics that each reference paper belongs to. We choose the topic
with the highest probability as the representative of a reference paper’s topic, assigning
each reference paper to only one topic.</p>
        <p>For ordering the sentences, we start with the topic of the paper with most citations,
adding the sentences from the papers that belongs to the same topic. Afterwards, we
repeat the process till all the papers have been included in the report.</p>
        <p>Tables 5 and 4 in the appendix show examples of generated related work reports with
and without topic modeling applied.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.4. Rewriting Models</title>
        <p>
          In our recent work [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ], we have used pointer–generator neural networks with two diferent
architectures; Bidirectional Recurrent Neural Networks (BRNN) [
          <xref ref-type="bibr" rid="ref29">29</xref>
          ] and Transformers
[
          <xref ref-type="bibr" rid="ref30">30</xref>
          ] to train a system able to generate citation-sentences from the title and abstract of a
research paper. The pointer–generator networks can copy words from the source text via
pointing, which aids accurate reproduction of information while retaining the ability to
produce novel words through the generator.
        </p>
        <p>
          In order to improve our extractive summaries, we use the selected sentences and
rephrase them using a set of pretrained abstractive models [
          <xref ref-type="bibr" rid="ref28">28</xref>
          ]. These models were
trained using over 16K pairs of Title and Abstract of scientific papers as a source sequence
(input) with a citation context as a target sequence (output), making them ideal for our
experiments. We feed the pretrained models with a scientific context extracted from
our extractive methods to generate a rephrased citation context that can represent a
scientific paper. Such additional experiments will avoid the use of copy-paste techniques
adopted by the extractive models. Once we have a citation context for each reference
paper we concatenate them into the final related work report.
        </p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments</title>
      <p>We use the Multi-level Annotated Corpus of Scientific Papers as our main data set
(testing data) aiming at recreating the related work section (report) of the target paper
for each cluster provided. We compared our systems against a set of of-the-shelf baselines.
We performed both automatic and human evaluations by comparing the systems to the
gold related work sections of the target paper in the Multi-level Annotated Corpus of
Scientific Papers.</p>
      <p>For scoring the sentences of the reference papers, we ran the unsupervised approaches:
Modified Jaccard and BabelNet Embeddings Distance directly over each cluster in the
test data. We modeled a pair of vectors from Level 2 of the corpus: the citation context
mentioned in the citing papers alongside each and every sentence in the corresponding
cited reference paper. Then we scored the sentences, that is, the cosine similarity score in
case of the BabelNet Embeddings and the score of the Modified Jaccard similarity. We
also trained several CNN models based on the similarity between the reference papers
sentences and the abstract, community and human written summaries provided by the
CL-Scisumm Challenge Dataset as a training dataset. We trained 6 models for each
summary of the reference paper each representing one of the scores: cosine distance:
based on ACL, Google and SUMMA, ROUGE-2 similarity and two averages of the four
scores: including and excluding SUMMA. After training our CNN models we fed them
each and every sentence from the reference paper in the testing data set (the multi-level
corpus). Once we scored the sentences of the reference papers using all our methods we
sort the sentences in each reference paper in descending order based on the score.</p>
      <p>After that, we select the top  sentences from each reference paper in a unified way
based on the length of the gold standard. Moreover, we organize the sentences of each
reference papers based on their topics. All reference papers that belongs to the same
topic were grouped together, the position of the sentences inside each reference paper
were respected. We have also generated related work reports without applying topic
modeling. Finally, we always generate related work reports that has the same number of
sentences as the gold related work report ( system =  gold).</p>
      <p>
        We perform additional experiments by selecting the top 1, 3, and 6 sentences from each
reference paper, passing them through a set of pretrained abstractive models to produce
citation contexts. Finally, we concatenate the citation contexts of the reference papers to
form the final related work report. Since [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ] produced several pretrained models based
on BRNN and Transformer architectures, we have used all of the pretrained models they
provided only reporting here the best models.
      </p>
      <sec id="sec-4-1">
        <title>4.1. Baselines</title>
        <p>
          For our experiments we implemented several extractive summarization baselines alongside
a set of simple baselines based on the observations arising from the analysis of citation
sentences and scientific abstracts on the use of titles and abstracts [
          <xref ref-type="bibr" rid="ref2 ref31">2, 31</xref>
          ]. The title
baseline is to use the title of each cited article as citation sentences. The abstract first
baseline uses as citation sentences the first sentence of the abstract of the cited articles
while the abstract last baseline uses the last sentence.
        </p>
        <p>
          The second set of baselines is composed of available systems that use well-established
extractive techniques. We have made sure that all the baselines have the same conditions
as our systems. That is we fed each and every scientific paper to the baseline and
guaranteed that at least the system will select one sentence from each. We also instructed
the system to generate the same number of sentences as the gold related work sections
( system =  gold). We describe the systems as follows:
• MEAD [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] is a well-known extractive document summarizer which generates
summaries using centroids alongside other features such as the position of the
sentence and the length.
• TextRank [
          <xref ref-type="bibr" rid="ref32">32</xref>
          ] and LexRank [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] are both extractive and unsupervised graph-based
text summarization systems which create sentence graphs in order to compute
centrality values for each sentence. Both algorithms have similar underlying methods
to compute centrality which are based on the PageRank ranking algorithm. They
difer in how links are weighted in the document graph.
• SUMMA [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ] is a Java implementation of several sentence scoring functions. We
use the implementation of the centroid scoring functionality to select the most
central sentence in a document.
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results, Evaluation and Discussion</title>
      <p>In this section we compare our systems against the baseline systems for the task of
automatic generation of related work reports. We have performed both automatic and
human evaluation. For automatic evaluation we have used 4 ROUGE metrics: ROUGE-1,
ROUGE-2, ROUGE-L and ROUGE-SU4. We present all the systems except for the CNN
approach for which we only present the top five systems. ROUGE measures combine
precision and recall in a harmonic F-measure which is generally used to assess the systems’
performance. The results of ROUGE-1 and ROUGE-2 metrics can be found in Table
1, while for ROUGE-L and ROUGE-SU4 the results are presented in Table 6 in the
appendix.</p>
      <p>The non-informed extractive baselines which do not perform any analysis of the input
(e.g. use of titles or sentences from abstracts) tend to have a high precision but low
recall, especially precise is the title. Except for LexRank, the of-the-shelf baselines
have low performance, which was expected since they are based on poor word-based
representations of the document. A CNN approach which uses word embedding, rich
summarization features, and scores sentences based on similarity to abstract outperforms
(in terms of ROUGE-1) all the other systems.</p>
      <p>
        Finally, we perform automatic evaluation using ROUGE of the rephrasing experiments
(only the top five best models are shown, all of them based on BRNN). We report the
results of selecting the top one, three and six sentences from the reference papers and
SYSTEM
Titles
AbsFS
AbsLS
SUMMA
MEAD
LexRank
TexRank
Babelnet
MJ
−2−
−
−
−2−
−2−ℎ
generating a citation context using the pretrained abstractive models trained by [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ].
The results are encouraging, the rewriting system improves the results of the extractive
methods in terms of ROUGE and also when comparing this with the pure abstractive
approach presented in [
        <xref ref-type="bibr" rid="ref28">28</xref>
        ], confirming that the selecting and rewriting approach to
summarization is a viable alternative to pure extractive or abstractive approaches.
      </p>
      <sec id="sec-5-1">
        <title>5.1. Human Evaluation</title>
        <p>In order to assess the quality of the automatically generated related work reports, we
selected 10 clusters that discusses diferent varieties of topics, each cluster was manually
evaluated by three subjects with the age group between 25-34 with an expert level of
English language, they have a range between good and very good in their expertise in
Natural Language Processing (the topic of the analyzed summaries).</p>
        <p>The objective of this evaluation is to assess the appropriateness of four diferent related
work sections for a given target paper in the test data set i.e. Multi-level corpus. The
four related work sections represent: the best system of the baselines i.e. LexRank, the
gold related work section and the best system we have with and without topic modeling
applied i.e.   − .</p>
        <p>To carry out the evaluation we prepared each reference paper’s Title, Abstract and
Introduction in PDF format. Alongside the scientific paper we provided in a random
order the related work sections in text format to be evaluated. We also added a folder
with the references that are mentioned in the related work section, we also provided the
bibliographic information about each of the references which will be cited in the related
work section. Given the target scientific paper Title, Abstract and Introduction section
alongside a related work section, we asked them for their opinion on three fronts:
• Responsiveness: How good do you consider the related work section given that it
must include information on the list of reference papers and must fit in the target
paper.
• Linguistic quality: How do you rate the readability and grammaticality of the
related work section? That is: is it understandable? is it grammatically correct
(are the sentences correct)? Are there any spelling mistakes? Is punctuation
appropriate?
• Text organization: How well organized and coherent the related work section is?
That is: does the discourse (topics) flows from sentences to sentence? Are the
sentences organized in a coherent way? Is the text not redundant?
We instructed them to read the target scientific paper’s Title, Abstract and Introduction
(the pdf file), and then to read each related work section (the text file). Once they had
ifnished reading the related work section we informed them to fill the evaluation form
indicating the scores for each metric, all the scores were on the scale of 1 to 5. Finally,
we requested that they should not check the web for a related work section or the target
paper to avoid influence from external variables and use the references folder if they felt
they had to.</p>
        <p>Table 3 present the average of all the metrics across the 10 clusters for our system
with and without topic modeling applied, LexRank: the best baseline in the automatic
evaluation and finally the gold related work report. What can be noticed is that our
system with topic modeling super-passes the baseline in all metrics and it is considered
an improvement over not implementing topic modeling for our system.</p>
        <sec id="sec-5-1-1">
          <title>System</title>
          <p>Gold
LexRank
WithoutTM
TopicModeling</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>In this paper, we have presented a number of computational approaches for the automatic
generation of related work reports. These approaches utilize extractive summarization
to describe each scientific paper to be mentioned in the related work report. We utilize
a corpus of articles connected through their citation network to test our approach. We
perform automatic and human evaluation over of our extractive methods showing their
viability. Moreover, we perform additional experiments to rephrase the sentences selected
by the extractive methods finding a viable alternative to pure extractive or abstractive
approaches. We have found that our regression learning extractive approach obtains
competitive results in terms of ROUGE scores when compared with well known baselines.
A human evaluation also confirms that the reports generated by the extractive approach
are also preferred to the best baseline. Moreover, automatically rewriting the sentences
which were automatically extracted produces abstracts which are better in terms of
ROUGE. We believe that ”select and rewrite” is a valid strategy to the generation of
summaries specially in the case of very long documents such as scientific articles. There
are many limitations of our work which indicate further research. In future work diferent
sentence ordering strategies should be investigated. Moreover, the generation of citation
sentences for multiple scientific papers is an interesting topic to take on.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>This work was (partly) supported by the Spanish Government under the María de
Maeztu Units of Excellence Programme (MDM-2015-0502). We also acknowledge support
from the project Context-aware Multilingual Text Simplification (ConMuTeS)
PID2019109066GB-I00/AEI/10.13039/501100011033 awarded by Ministerio de Ciencia, Innovación
y Universidades (MCIU) and by Agencia Estatal de Investigación (AEI) of Spain.</p>
      <sec id="sec-7-1">
        <title>Computational Linguistics, 2004.</title>
        <p>A. Appendix
(Blunsom et al. 2007) Statistical machine translation (SMT) has seen a resurgence in popularity
in recent years ... (Kumar and Byrne 2004) We also show how MBR decoding can be used to
incorporate syntactic structure into a statistical MT system ... template model for statistical
machine translation. (Matsusaki et al. 2005) This paper defines a generative probabilistic model
of parse trees, which we call PCFG-LA. This paper defines a generative model of parse trees that
we call PCFG with latent annotations (PCFG-LA). (May and Knight 2006) We also demonstrate
our algorithm’s efectiveness ... to deal with grammars that produce trees. (Petrov et al. 2006) In
this paper, we investigate the learning of a grammar consistent with a treebank at ... likelihood
of the training trees. We present a method that combines the strengths of both manual and
automatic approaches while addressing some of their common shortcomings. (Tromble et al.
2008) In this paper we explore a diferent strategy to perform MBR decoding over Translation
Lattices ... that compactly encode a huge number of translation ... We begin with a review of
MBR decoding for Statistical Machine Translation (SMT).
(Blunsom et al. 2007) Statistical machine translation (SMT) has seen a resurgence in popularity
in recent years ... (Kumar and Byrne 2004) We also show how MBR decoding can be used to
incorporate syntactic structure into a statistical MT system ... template model for statistical
machine translation. (Tromble et al. 2008) In this paper we explore a diferent strategy to
perform MBR decoding over Translation Lattices ... that compactly encode a huge number of
translation ... We begin with a review of MBR decoding for Statistical Machine Translation
(SMT). (Matsusaki et al. 2005) This paper defines a generative probabilistic model of parse
trees, which we call PCFG-LA. This paper defines a generative model of parse trees that we call
PCFG with latent annotations (PCFG-LA). (May and Knight 2006) We also demonstrate our
algorithm’s efectiveness ... to deal with grammars that produce trees. (Petrov et al. 2006) In
this paper, we investigate the learning of a grammar consistent with a treebank at ... likelihood
of the training trees. We present a method that combines the strengths of both manual and
automatic approaches while addressing some of their common shortcomings.
SYSTEM
Titles
AbsFS
AbsLS
SUMMA
MEAD
LexRank
TexRank
Babelnet
MJ
 −2−
−
−
 −2−
 −2−ℎ
1
3
6</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>C. S. G.</given-names>
            <surname>Khoo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.-C.</given-names>
            <surname>Na</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Jaidka</surname>
          </string-name>
          ,
          <article-title>Analysis of the macro-level discourse structure of literature reviews</article-title>
          .,
          <source>Online Information Review</source>
          <volume>35</volume>
          (
          <year>2011</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>K.</given-names>
            <surname>Jaidka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Khoo</surname>
          </string-name>
          , J.-C. Na,
          <article-title>Deconstructing human literature reviews-a framework for multi-document summarization</article-title>
          ,
          <source>in: Proceedings of the 14th European Workshop on Natural Language Generation</source>
          ,
          <year>2013</year>
          , pp.
          <fpage>125</fpage>
          -
          <lpage>135</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>C. D. V.</given-names>
            <surname>Hoang</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kan</surname>
          </string-name>
          ,
          <article-title>Towards automated related work summarization</article-title>
          ,
          <source>in: Proceedings of the 23rd International Conference on Computational Linguistics: Posters, Association for Computational Linguistics</source>
          ,
          <year>2010</year>
          , pp.
          <fpage>427</fpage>
          -
          <lpage>435</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wan</surname>
          </string-name>
          ,
          <article-title>Automatic generation of related work sections in scientific papers: An optimization approach</article-title>
          ., in: EMNLP,
          <year>2014</year>
          , pp.
          <fpage>1624</fpage>
          -
          <lpage>1633</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>N.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Gvr</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. S.</given-names>
            <surname>Reddy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. P.</given-names>
            <surname>Rosé</surname>
          </string-name>
          ,
          <article-title>Towards multi-document summarization of scientific articles: making interesting comparisons with scisumm</article-title>
          ,
          <source>in: Proceedings of the Workshop on Automatic Summarization for Diferent Genres, Media, and Languages, Association for Computational Linguistics</source>
          ,
          <year>2011</year>
          , pp.
          <fpage>8</fpage>
          -
          <lpage>15</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>H. P.</given-names>
            <surname>Luhn</surname>
          </string-name>
          ,
          <article-title>The automatic creation of literature abstracts</article-title>
          ,
          <source>IBM J. Res. Dev</source>
          .
          <volume>2</volume>
          (
          <year>1958</year>
          )
          <fpage>159</fpage>
          -
          <lpage>165</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>H. P.</given-names>
            <surname>Edmundson</surname>
          </string-name>
          ,
          <article-title>New methods in automatic extracting</article-title>
          ,
          <source>J. ACM</source>
          <volume>16</volume>
          (
          <year>1969</year>
          )
          <fpage>264</fpage>
          -
          <lpage>285</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>H.</given-names>
            <surname>Saggion</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Poibeau</surname>
          </string-name>
          ,
          <article-title>Automatic text summarization: Past, present andfuture</article-title>
          , in: T. Poibeau,
          <string-name>
            <given-names>H.</given-names>
            <surname>Saggion</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Piskorski</surname>
          </string-name>
          , R. Yangarber (Eds.), Multi-source,
          <source>Multilingual Information Extraction and Summarization</source>
          , Springer Verlag, Berlin,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>V.</given-names>
            <surname>Qazvinian</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. R.</given-names>
            <surname>Radev</surname>
          </string-name>
          ,
          <article-title>Scientific paper summarization using citation summary networks</article-title>
          ,
          <source>in: Proceedings of the 22nd International Conference on Computational Linguistics-Volume</source>
          <volume>1</volume>
          , Association for Computational Linguistics,
          <year>2008</year>
          , pp.
          <fpage>689</fpage>
          -
          <lpage>696</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Cohan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goharian</surname>
          </string-name>
          ,
          <article-title>Scientific article summarization using citation-context and article's discourse structure</article-title>
          ,
          <source>arXiv preprint arXiv:1704.06619</source>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>X.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhuge</surname>
          </string-name>
          ,
          <article-title>Summarization of scientific paper through reinforcement ranking on semantic link network</article-title>
          ,
          <source>IEEE Access 6</source>
          (
          <year>2018</year>
          )
          <fpage>40611</fpage>
          -
          <lpage>40625</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Yao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <article-title>Event-based summarization method for scientific literature</article-title>
          ,
          <source>Personal and Ubiquitous Computing</source>
          (
          <year>2019</year>
          )
          <fpage>1</fpage>
          -
          <lpage>10</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>H. C. D. Vu</surname>
          </string-name>
          ,
          <source>Towards automated related work summarization</source>
          ,
          <source>Ph.D. thesis</source>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>T.</given-names>
            <surname>Hofmann</surname>
          </string-name>
          ,
          <article-title>Probabilistic latent semantic indexing</article-title>
          ,
          <source>in: Proceedings of the 22nd annual international ACM SIGIR conference on Research and development in information retrieval, ACM</source>
          ,
          <year>1999</year>
          , pp.
          <fpage>50</fpage>
          -
          <lpage>57</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>D. R.</given-names>
            <surname>Radev</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Allison</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Blair-Goldensohn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Blitzer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Celebi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dimitrov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Drabek</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hakim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Lam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Liu</surname>
          </string-name>
          , et al.,
          <article-title>Mead-a platform for multidocument multilingual text summarization</article-title>
          .,
          <source>in: LREC</source>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>G.</given-names>
            <surname>Erkan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. R.</given-names>
            <surname>Radev</surname>
          </string-name>
          , Lexrank:
          <article-title>Graph-based lexical centrality as salience in text summarization</article-title>
          ,
          <source>Journal of Artificial Intelligence Research</source>
          <volume>22</volume>
          (
          <year>2004</year>
          )
          <fpage>457</fpage>
          -
          <lpage>479</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>W.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Itoh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tanida</surname>
          </string-name>
          ,
          <string-name>
            <surname>Y.</surname>
          </string-name>
          <article-title>Ichioka, Parallel distributed processing model with local space-invariant interconnections and its optical architecture</article-title>
          ,
          <source>Applied optics 29</source>
          (
          <year>1990</year>
          )
          <fpage>4790</fpage>
          -
          <lpage>4797</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <surname>D. M. Blei</surname>
            ,
            <given-names>A. Y.</given-names>
          </string-name>
          <string-name>
            <surname>Ng</surname>
            ,
            <given-names>M. I. Jordan</given-names>
          </string-name>
          ,
          <article-title>Latent dirichlet allocation</article-title>
          ,
          <source>Journal of machine Learning research 3</source>
          (
          <year>2003</year>
          )
          <fpage>993</fpage>
          -
          <lpage>1022</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19] A. AbuRa'ed,
          <string-name>
            <given-names>H.</given-names>
            <surname>Saggion</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Chiruzzo</surname>
          </string-name>
          ,
          <article-title>A multi-level annotated corpus of scientific papers for scientific document summarization and cross-document relation discovery</article-title>
          ,
          <source>in: Proceedings of The 12th Language Resources and Evaluation Conference</source>
          , European Language Resources Association, Marseille, France,
          <year>2020</year>
          , pp.
          <fpage>6672</fpage>
          -
          <lpage>6679</lpage>
          . URL: https://www.aclweb.org/anthology/2020.lrec-
          <volume>1</volume>
          .
          <fpage>824</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>R.</given-names>
            <surname>Navigli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. P.</given-names>
            <surname>Ponzetto</surname>
          </string-name>
          ,
          <article-title>BabelNet: The automatic construction, evaluation and application of a wide-coverage multilingual semantic network</article-title>
          ,
          <source>Artificial Intelligence</source>
          <volume>193</volume>
          (
          <year>2012</year>
          )
          <fpage>217</fpage>
          -
          <lpage>250</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>S.</given-names>
            <surname>Niwattanakul</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Singthongchai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Naenudorn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Wanapu</surname>
          </string-name>
          ,
          <article-title>Using of jaccard coefifcient for keywords similarity, in: Proceedings of the international multiconference of engineers and computer scientists</article-title>
          , volume
          <volume>1</volume>
          ,
          <year>2013</year>
          , pp.
          <fpage>380</fpage>
          -
          <lpage>384</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>M.</given-names>
            <surname>Mancini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Camacho-Collados</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Iacobacci</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Navigli</surname>
          </string-name>
          ,
          <article-title>Embedding words and senses together via joint knowledge-enhanced training</article-title>
          ,
          <source>arXiv preprint arXiv:1612.02703</source>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <surname>M. K. Chandrasekaran</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Yasunaga</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Radev</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Freitag</surname>
            , M.-
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>Kan</surname>
          </string-name>
          ,
          <article-title>Overview and results: Cl-scisumm shared task 2019</article-title>
          , arXiv preprint arXiv:
          <year>1907</year>
          .
          <volume>09854</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>H.</given-names>
            <surname>Saggion</surname>
          </string-name>
          , A. AbuRa'ed,
          <string-name>
            <given-names>F.</given-names>
            <surname>Ronzano</surname>
          </string-name>
          ,
          <article-title>Trainable citation-enhanced summarization of scientific articles</article-title>
          , in: G. Cabanac,
          <string-name>
            <given-names>M. K.</given-names>
            <surname>Chandrasekaran</surname>
          </string-name>
          , I. Frommholz,
          <string-name>
            <given-names>K.</given-names>
            <surname>Jaidka</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Kan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mayr</surname>
          </string-name>
          , D. Wolfram (Eds.),
          <source>Proceedings of the Joint Workshop on Bibliometric-enhanced Information Retrieval</source>
          and
          <article-title>Natural Language Processing for Digital Libraries (BIRNDL) co-located with the Joint Conference on Digital Libraries 2016 (JCDL 2016), Newark</article-title>
          , NJ, USA, June 23,
          <year>2016</year>
          , volume
          <volume>1610</volume>
          <source>of CEUR Workshop Proceedings, CEUR-WS.org</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>175</fpage>
          -
          <lpage>186</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>H.</given-names>
            <surname>Saggion</surname>
          </string-name>
          ,
          <string-name>
            <surname>SUMMA.</surname>
          </string-name>
          <article-title>A Robust and Adaptable Summarization Tool</article-title>
          , TAL
          <volume>49</volume>
          (
          <year>2008</year>
          )
          <fpage>103</fpage>
          -
          <lpage>125</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <surname>C.-Y. Lin</surname>
            ,
            <given-names>ROUGE:</given-names>
          </string-name>
          <article-title>A package for automatic evaluation of summaries, in: Text summarization branches out:</article-title>
          <source>Proceedings of the ACL-04 workshop</source>
          , volume
          <volume>8</volume>
          ,
          <string-name>
            <surname>Barcelona</surname>
          </string-name>
          , Spain,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27] A. AbuRa'ed, À. Bravo
          <string-name>
            <surname>Serrano</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Chiruzzo</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Saggion</surname>
          </string-name>
          , Lastus/taln+ inco@
          <article-title>cl-scisumm 2018-using regression and convolutions for cross-document semantic linking and summarization of scholarly literature</article-title>
          , in: Mayr P,
          <string-name>
            <surname>Chandrasekaran</surname>
            <given-names>MK</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jaidka</surname>
            <given-names>K</given-names>
          </string-name>
          , editors.
          <source>BIRNDL 2018. 3rd Joint Workshop on Bibliometric-enhanced Information Retrieval and Natural Language Processing for Digital Libraries; 2018 Jul</source>
          <volume>21</volume>
          ; Ann Arbor, MI.[place unknown]: CEUR;
          <year>2018</year>
          . p.
          <fpage>150</fpage>
          -
          <lpage>63</lpage>
          ., CEUR Workshop Proceedings,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28] A. AbuRa'ed,
          <string-name>
            <given-names>H.</given-names>
            <surname>Saggion</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Shvets</surname>
          </string-name>
          , À. Bravo,
          <article-title>Automatic related work section generation: experiments in scientific document abstracting</article-title>
          ,
          <source>Scientometrics</source>
          <volume>125</volume>
          (
          <year>2020</year>
          )
          <fpage>3159</fpage>
          -
          <lpage>3185</lpage>
          . URL: https://doi.org/10.1007/s11192-020-03630-2. doi:
          <volume>10</volume>
          .1007/ s11192-020-03630-2.
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>M.</given-names>
            <surname>Schuster</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Paliwal</surname>
          </string-name>
          ,
          <article-title>Bidirectional recurrent neural networks</article-title>
          ,
          <source>IEEE Transactions on Signal Processing</source>
          <volume>45</volume>
          (
          <year>1997</year>
          )
          <fpage>2673</fpage>
          -
          <lpage>2681</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>A.</given-names>
            <surname>Vaswani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Parmar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Uszkoreit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Jones</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. N.</given-names>
            <surname>Gomez</surname>
          </string-name>
          , Ł. Kaiser,
          <string-name>
            <surname>I. Polosukhin</surname>
          </string-name>
          ,
          <article-title>Attention is all you need</article-title>
          ,
          <source>in: Advances in neural information processing systems</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>5998</fpage>
          -
          <lpage>6008</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>H.</given-names>
            <surname>Saggion</surname>
          </string-name>
          ,
          <article-title>Using linguistic knowledge in automatic abstracting, in: 27th Annual Meeting of the Association for Computational Linguistics</article-title>
          , University of Maryland, College Park, Maryland, USA,
          <fpage>20</fpage>
          -
          <lpage>26</lpage>
          June 1999.,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>R.</given-names>
            <surname>Mihalcea</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Tarau</surname>
          </string-name>
          , Textrank: Bringing order into texts, Association for
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>