<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Leveraging Transformers for Structured Highlight Generation in Scientific Literature</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Rachana Nagaraju</string-name>
          <email>rachananagaraju20@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Hosahalli Lakshmaiah Shashirekha</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, Mangalore University</institution>
          ,
          <addr-line>Mangalore, Karnataka</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2026</year>
      </pub-date>
      <abstract>
        <p>The rapid expansion of scientific literature has made it increasingly dificult for researchers to keep track of key contributions across disciplines. While abstracts provide useful overviews, they are often too detailed for quick comprehension. Research highlights address this gap by presenting concise, structured statements that emphasize the most important findings of a paper. Automating the generation of such highlights has the potential to accelerate knowledge discovery, improve the accessibility of research outputs, and support academic search engines and digital libraries in indexing scientific content more efectively. The SciHigh shared task is designed to explore the challenges of generating highlights from abstracts. In this paper, we, Team MUCS, present our transformer-based summarization pipeline built on the T5-base model, developed and submitted as part of the SciHigh shared task. The pipeline incorporated pre-processing, tokenization, and fine-tuning within a sequence-to-sequence (seq2seq) framework. Training is carried out using the AdamW optimizer with careful parameter selection, while generation employed constrained decoding strategies to balance informativeness and lfuency. The design emphasized capturing both the lexical and semantic correspondence between abstracts and their highlights, with evaluation relying on Recall-Oriented Understudy for Gisting Evaluation (ROUGE) - a widely used metrics for summarization. Our system achieved a ROUGE-L score of 0.2208, placing us at 7th rank on the oficial leaderboard. These results highlight the strength of fine-tuned transformer architectures for scientific highlight generation, while also revealing opportunities for further improvement through the integration of retrieval-augmented approaches, hybrid architectures, or domain-adaptive pretraining. Overall, our study demonstrates how modern neural summarization techniques can be efectively applied to generate structured highlights, ultimately contributing to more eficient scientific communication and knowledge dissemination.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Research Highlight Generation</kwd>
        <kwd>Scientific Summarization</kwd>
        <kwd>Transformer Models</kwd>
        <kwd>T5</kwd>
        <kwd>Sequence-to-Sequence model</kwd>
        <kwd>Automatic Text Summarization</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Digital repositories and online journals publish thousands of articles daily, making eficient mechanisms
for information retrieval and comprehension essential [
        <xref ref-type="bibr" rid="ref1 ref2">1, 2</xref>
        ]. But the unprecedented growth of scientific
literature in recent years has made it increasingly dificult for researchers to stay updated with key
ifndings across domains. Traditionally, abstracts serve as the primary medium for summarizing research
papers. While they provide valuable overviews, abstracts are often lengthy and written in prose, making
them less accessible for quick scanning, particularly on mobile devices or in time-sensitive scenarios.
      </p>
      <p>
        Research highlights are introduced by publishers as a complementary form of scientific summarization.
Unlike abstracts, highlights are short, bullet-style statements that emphasize the most significant
contributions of a study. They have been shown to aid rapid comprehension, improve discoverability
in academic search engines, and enhance metadata quality for digital libraries [
        <xref ref-type="bibr" rid="ref3 ref4">3, 4</xref>
        ]. Automating the
generation of such highlights can significantly reduce the cognitive load on researchers and streamline
the scientific communication process. The diference between abstract and highlights is illustrated
in Table 1. It can be observed that highlights provide concise and accessible summaries compared to
longer and more detailed abstracts.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Works</title>
      <p>
        Generation of highlights from scientific articles has received increasing attention as a subdomain
of summarization, particularly motivated by the need to produce concise and domain-relevant
summaries of research contributions. Various approaches have explored this task ranging from extractive
models to pre-trained transformer architectures and hybrid methods. Rehman et al. [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] proposed a
pointer-generator network enhanced with a coverage mechanism and SciBERT embeddings, specifically
targeting highlight generation from scientific abstracts. Their model achieved ROUGE-1, ROUGE-2,
and ROUGE-L scores of 41.78, 18.63, and 31.46 respectively on the CSPubSum dataset and 42.13, 18.91,
and 31.58 on MixSub, outperforming traditional baselines. The system demonstrated strong handling of
1https://sites.google.com/jadavpuruniversity.in/scihigh2025
2https://fire.irsi.org.in/fire/2025/home
3https://github.com/rachanabn20/SciHigh-2025
inherent scientific terminology and reduced redundancy during generation. However, the reliance on
pointer networks introduced an extractive bias, preventing truly abstractive and novel rephrasings. In a
related study, Rehman et al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] explored the use of ELMo contextual embeddings to enrich the
semantic representations in highlight generation models. While the method improved word-level context
understanding and achieved competitive scores (ROUGE-L of 26.42 on CSPubSum), its summarization
quality lagged behind transformer-based models, largely due to ELMo’s limited capacity to handle long
dependencies and capture sentence-level discourse structure.
      </p>
      <p>
        Rehman et al. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] further introduced an Named Entity Recognition (NER)-augmented summarization
pipeline, where named entity information was integrated during the generation process. This technique
improved content relevance and model interpretability and was shown to increase ROUGE-1 and
METEOR scores by approximately 1.5–2 points over non-NER baselines. However, its efectiveness was
heavily dependent on the accuracy of the underlying NER module, particularly for domain-specific
entities. In a benchmarking study, Rehman et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] evaluated several pre-trained summarization
models and analyzed their performance on scientific text. The study reported that models such as T5
and BART struggled with domain mismatch unless fine-tuned properly, and models pre-trained on
newswire often failed to generalize to scientific abstracts. Rehman et al. [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] also proposed a hybrid
architecture combining extractive components with neural abstractive summarization. Though efective
at maintaining key phrases from abstracts, the methodology lacked deep abstraction and was sensitive
to input noise and alignment mismatches.
      </p>
      <p>
        Xiao et al. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] proposed PRIMERA, a pyramid-based pretraining strategy optimized for long and
multidocument scientific summarization. Their model achieved impressive scores, reporting ROUGE-1/2/L as
47.2/20.8/39.6 on long document datasets, and outperformed BART and PEGASUS in terms of
sentencelevel coherence. PRIMERA explicitly models sentence dependencies and performs well in scenarios
requiring reasoning over multiple sentences, but it comes at the cost of significant computational
overhead and susceptibility to hallucination, particularly in domain-specific tasks. Lewis et al. [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ]
presented Retrieval-Augmented Generation (RAG), a hybrid architecture that pairs dense passage
retrievers with generative models to inject factual information into generated summaries. This approach
demonstrated improved factual accuracy in knowledge-intensive tasks but posed engineering challenges
due to its dependency on retrieval quality and alignment with the generator. Further, performance
often degraded when the retriever failed to retrieve relevant passages, especially in unseen domains.
      </p>
      <p>
        Zhang et al. [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] introduced PEGASUS, a summarization-oriented pretraining framework where
key sentences are masked and reconstructed. PEGASUS achieved state-of-the-art ROUGE-L scores
(45.16 on scientific summarization benchmarks) and aligned closely with actual summarization tasks.
Its pretraining objective significantly boosted performance in low-resource scenarios. However, the
model’s efectiveness was reliant on high-quality sentence masking heuristics, and its training demands
were computationally expensive. Lewis et al. [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], in contrast, proposed BART, a denoising autoencoder
that is commonly fine-tuned for summarization tasks. BART has become a popular baseline for scientific
summarization due to its robustness and fluency, achieving a ROUGE-L of up to 36.93 on the arXiv
dataset. However, like many generative models, it sometimes hallucinates facts and lacks grounding in
the input abstract.
      </p>
      <p>
        La Quatra et al. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] developed THExt, a highlights extraction model trained on author-provided
highlights. Their system relied on selecting sentences based on transformer-encoded importance
scores. It ofered eficiency and simplicity but was inherently extractive, thus incapable of performing
paraphrasing or abstraction beyond sentence boundaries. Goyal et al. [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] explored few-shot and
zero-shot summarization using prompt-based large language models. These models, including GPT-3
and its successors, demonstrated strong fluency and generalization ability. Without fine-tuning, they
generated highlights of competitive quality. However, they require extensive computational resources
and often hallucinate scientific terminology or infer conclusions not in the source text.
      </p>
      <p>
        In summary, highlight generation techniques range from lightweight, domain-trained models [
        <xref ref-type="bibr" rid="ref5 ref6 ref7">5, 6, 7</xref>
        ]
to large, general-purpose summarization frameworks [
        <xref ref-type="bibr" rid="ref10 ref12 ref13">12, 13, 10</xref>
        ]. Models such as PEGASUS and
PRIMERA ofer strong abstractive capacity and high ROUGE scores, but are computationally expensive
and somewhat opaque. Pointer-generator and NER-augmented methods provide more interpretable and
domain-aligned summaries while remaining lightweight, though they often trend toward extractiveness.
Hybrid and retrieval-based models like RAG improve factual accuracy but struggle with complexity.
Prompt-based LLMs show strong zero-shot potential and fluency but face issues with hallucination
and cost. The extensive work by Rehman et al. across multiple architectures highlights the need
for cost-efective and interpretable models tailored specifically to scientific summarization, balancing
performance and practicality.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>The methodology for the SciHigh shared task is designed around a transformer-based summarization
pipeline. In our approach, the abstracts are first pre-processed to ensure consistency in length and format.
These pre-processed texts are then transformed into feature representations using the Text-to-Text
Transfer Transformer (T5)-base4 tokenizer, which encodes both the abstracts and their corresponding
highlights.</p>
      <p>Finally, a seq2seq model based on T5-base is fine-tuned on the training data to map abstracts into
concise, highlight-style statements.</p>
      <p>T5 model is a pre-trained seq2seq architecture introduced by Google Research. Built on the
Transformer architecture, T5 is trained on a large corpus known as the C4 dataset, allowing it to learn
powerful language representations. It treats every Natural Language Processing (NLP) task as a
text-totext problem, enabling a unified approach to tasks like translation, summarization, and classification.
The T5-base variant used in our work contains 220 million parameters, balancing performance and
computational eficiency.</p>
      <p>The proposed pipeline enables the system to learn the structural diferences between abstracts
and highlights, ultimately producing outputs that resemble author-written highlights. The overall
architecture of our proposed retrieval framework is illustrated in Figure 1.</p>
      <sec id="sec-3-1">
        <title>3.1. Text Pre-processing</title>
        <p>
          MixSub-SciHigh dataset - a subset of MixSub dataset [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] provided for SciHigh shared task contains
research articles collected from ScienceDirect across diferent scientific domains, published in 2020.
Each instance has an abstract along with the corresponding author-written research highlights. Since
transformer models require uniform tokenized inputs, the following pre-processing steps are applied to
ensure consistency:
• Lowercasing and whitespace normalization: All text is converted to lowercase and
unnecessary whitespace characters are stripped of.
• Input length standardization: Abstracts are tokenized using the T5 tokenizer and truncated or
padded to a maximum length of 512 tokens to fit the model’s input constraints.
• Target length standardization: Abstracts are tokenized and truncated or padded to 512 tokens,
while research highlights are standardized to 128 tokens.
4https://huggingface.co/docs/transformers/en/model_doc/t5
• Special tokens: Padding tokens and sentence delimiters (&lt;pad&gt;, &lt;/s&gt;) are preserved to maintain
sequence boundaries.
        </p>
        <p>This pre-processing ensures that both abstracts and highlights are consistently represented as
fixedlength sequences suitable for transformer input.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Feature Representation</title>
        <p>After pre-processing, textual inputs are transformed into numerical features using the AutoTokenizer
from HuggingFace:
• Input IDs: Each abstract is mapped into a sequence of integer token IDs representing words and
sub-words.
• Attention masks: Binary masks are generated to distinguish between padded tokens and valid
tokens. This prevents the model from attending to padded positions during training.
• Labels: Highlights are similarly tokenized and mapped to integer IDs, which serve as ground-truth
labels during model training.</p>
        <p>These feature representations capture the semantic information of abstracts while maintaining structural
alignment with highlights.
3.3. Model Training
seq2seq models are widely used for text generation tasks, as they learn to transform an input sequence
into a corresponding output sequence. We fine-tune a pre-trained T5-base5 model using the
HuggingFace AutoModelForSeq2SeqLM API to map abstracts (source sequence) into highlights (target
sequence) using the following steps:
• Batching: Training and Validation sets are divided into mini-batches for eficient GPU utilization.
• Forward pass: Each batch of inputs (input IDs, attention masks, labels) is passed into the model.</p>
        <p>The decoder generates highlights corresponding to the abstract.
• Loss computation: Cross-entropy loss is computed between predicted sequences and reference
highlights.
• Optimization: The AdamW optimizer updates model parameters, with weight decay to prevent
overfitting.
• Gradient management: Gradients are reset after every batch using optimizer.zero_grad()
to ensure stability.
• Epochs: The model is trained for two epochs, with validation performed after each epoch to
track ROUGE scores.</p>
        <p>The hyperparameters used in training the model are listed in Table 2.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments and Results</title>
      <p>5https://huggingface.co/t5-base
6https://pytorch.org/
7https://huggingface.co/docs/transformers/index
8https://huggingface.co/docs/evaluate
The implementation relies on PyTorch6 for Deep Learning, HuggingFace Transformers7 for seq2seq
modeling, and the Evaluate library8 for computing metrics. Details of the dataset are shown in Table 3.
The models are evaluated based on ROUGE-1, ROUGE-2, ROUGE-L, and ROUGE-Lsum scores. ROUGE
is a widely used metric for summarization tasks. While ROUGE-1 measures unigram overlap between
generated and reference highlights, reflecting basic content coverage, ROUGE-2 captures bigram
overlap, providing insights into fluency and phrase-level accuracy. ROUGE-L evaluates the longest
common subsequence (LCS), highlighting the ability of the system to preserve structural consistency
and ROUGE-Lsum is a summary-level variant of ROUGE-L, designed to evaluate matches across
multiple sentences, making it particularly suitable for highlight-style summarization. These metrics
are chosen since they are widely adopted in automatic text summarization benchmarks, including the
SciHigh shared task at FIRE 2025.</p>
      <sec id="sec-4-1">
        <title>4.1. Results</title>
        <p>ROUGE scores obtained by the prosposed models on Validation and Test sets are shown in Table 4. Our
proposed model obtained a ROUGE-L score of 0.2208 on the Test set with 7th rank. The results illustrate
that the model performed reasonably well on Validation set, with ROUGE-1 score 0.3333 and ROUGE-L
score 0.2442. On the Test set, there is a small drop (ROUGE-1: 0.2961, ROUGE-L: 0.2208), showing
that while the model generalizes, it struggles a bit with unseen abstracts from diferent domains.
Interestingly, the ROUGE-Lsum score on the Test set (0.2500) is slightly higher than on the Validation set
(0.2444). This suggests that the generated highlights capture the sentence-level summary structure fairly
well. Overall, the fine-tuned t5-base model can produce concise and structured highlights, but there
is still room to improve through domain adaptation, longer training, or combining with retrieval-based
methods. Figure 2 shows the performance of the all the participating teams in the shared task.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion and Future Work</title>
      <p>In this work, we describe the model submitted to SciHigh shared task at FIRE 2025 on research highlight
generation using abstracts. Our approach based on fine-tuning the t5-base model achieved a ROUGE-L
score of 0.2208 on the oficial test set, placing our team- MUCS at the 7 th rank on the leaderboard.
The results indicate that transformer-based summarization models are efective in generating concise
and structured highlights, though some performance gaps remain compared to the top-ranked systems.
Eforts towards improving generalization across diverse scientific domains and exploring lightweight
adaptations to better capture domain-specific writing styles will be explored further. Such enhancements
could further improve the quality and readability of automatically generated highlights.</p>
    </sec>
    <sec id="sec-6">
      <title>6. Declaration on Generative AI</title>
      <p>Generative AI tools are used in assisting with language refinement and formatting of certain sections of
this paper. All technical content, experiments, results, and interpretations are conceived, implemented,
and validated independently. The AI tool does not contribute to the design of experiments, execution of
code, data analysis, or interpretation of results. Final responsibility for the accuracy and integrity of the
paper remains with the contributors.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>L.</given-names>
            <surname>Bornmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Mutz</surname>
          </string-name>
          ,
          <article-title>Growth Rates of Modern Science: A Bibliometric Analysis Based on the Number of Publications and Cited References</article-title>
          ,
          <source>Journal of the Association for Information Science and Technology</source>
          <volume>66</volume>
          (
          <year>2015</year>
          )
          <fpage>2215</fpage>
          -
          <lpage>2222</lpage>
          . doi:
          <volume>10</volume>
          .1002/asi.23329.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>Fortunato</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. T.</given-names>
            <surname>Bergstrom</surname>
          </string-name>
          , et al.,
          <source>Science of Science, Science</source>
          <volume>359</volume>
          (
          <year>2018</year>
          )
          <article-title>eaao0185</article-title>
          . doi:
          <volume>10</volume>
          .1126/ science.aao0185.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>P.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Lam</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          <article-title>Survey on Abstractive Text Summarization</article-title>
          ,
          <source>Proceedings of the AAAI Conference on Artificial Intelligence</source>
          <volume>34</volume>
          (
          <year>2020</year>
          )
          <fpage>9380</fpage>
          -
          <lpage>9387</lpage>
          . doi:
          <volume>10</volume>
          .1609/aaai.v34i05.
          <fpage>6396</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>R.</given-names>
            <surname>Nallapati</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Gulcehre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Xiang</surname>
          </string-name>
          , et al.,
          <article-title>Abstractive Text Summarization Using Sequenceto-Sequence RNNs and Beyond</article-title>
          ,
          <source>in: Proceedings of the 20th SIGNLL Conference on Computational Natural Language Learning (CoNLL)</source>
          ,
          <year>2016</year>
          , pp.
          <fpage>280</fpage>
          -
          <lpage>290</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>K16</fpage>
          -1028.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>T.</given-names>
            <surname>Rehman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. K.</given-names>
            <surname>Sanyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chattopadhyay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Bhowmick</surname>
          </string-name>
          ,
          <string-name>
            <surname>P. P. Das</surname>
          </string-name>
          ,
          <article-title>Generation of highlights from research papers using pointer-generator networks and SciBERT embeddings</article-title>
          ,
          <source>IEEE Access 11</source>
          (
          <year>2023</year>
          )
          <fpage>91358</fpage>
          -
          <lpage>91374</lpage>
          . doi:
          <volume>10</volume>
          .1109/ACCESS.
          <year>2023</year>
          .
          <volume>3292300</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Rehman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. K.</given-names>
            <surname>Sanyal</surname>
          </string-name>
          , S. Chattopadhyay,
          <article-title>Research highlight generation with ELMo contextual embeddings</article-title>
          ,
          <source>Scalable Computing: Practice and Experience</source>
          <volume>24</volume>
          (
          <year>2023</year>
          )
          <fpage>181</fpage>
          -
          <lpage>190</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>T.</given-names>
            <surname>Rehman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. K.</given-names>
            <surname>Sanyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Majumder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chattopadhyay</surname>
          </string-name>
          ,
          <article-title>Named entity recognition based automatic generation of research highlights</article-title>
          ,
          <source>in: Proceedings of the Third Workshop on Scholarly Document Processing (SDP</source>
          <year>2022</year>
          )
          <article-title>collocated with COLING 2022, Association for Computational Linguistics</article-title>
          , Gyeongju, Republic of Korea,
          <year>2022</year>
          , pp.
          <fpage>163</fpage>
          -
          <lpage>169</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>T.</given-names>
            <surname>Rehman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Das</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. K.</given-names>
            <surname>Sanyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chattopadhyay</surname>
          </string-name>
          ,
          <article-title>An analysis of abstractive text summarization using pre-trained models</article-title>
          ,
          <source>in: Proceedings of International Conference on Computational Intelligence, Data Science and Cloud Computing: IEM-ICDC 2021</source>
          , Springer,
          <year>2022</year>
          , pp.
          <fpage>253</fpage>
          -
          <lpage>264</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>T.</given-names>
            <surname>Rehman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. K.</given-names>
            <surname>Sanyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chattopadhyay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Bhowmick</surname>
          </string-name>
          ,
          <string-name>
            <surname>P. P. Das</surname>
          </string-name>
          ,
          <article-title>Automatic generation of research highlights from scientific abstracts</article-title>
          , in: 2nd Workshop on Extraction and
          <article-title>Evaluation of Knowledge Entities from Scientific Documents (EEKE'21), collocated with</article-title>
          <source>JCDL'21</source>
          ,
          <year>2021</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>W.</given-names>
            <surname>Xiao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Carenini</surname>
          </string-name>
          , et al.,
          <article-title>PRIMERA: Pyramid-based Masked Sentence Pre-training for Multidocument Summarization</article-title>
          ,
          <source>in: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (ACL)</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>7893</fpage>
          -
          <lpage>7909</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2022</year>
          .
          <article-title>acl-long</article-title>
          .
          <volume>542</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>P.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Perez</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Piktus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Petroni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Karpukhin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          , et al.,
          <article-title>Retrieval-augmented Generation for Knowledge-Intensive NLP</article-title>
          ,
          <source>in: Advances in Neural Information Processing Systems (NeurIPS)</source>
          , volume
          <volume>33</volume>
          ,
          <year>2020</year>
          , pp.
          <fpage>9459</fpage>
          -
          <lpage>9474</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Saleh</surname>
          </string-name>
          , P. J. Liu, PEGASUS:
          <article-title>Pre-training with Extracted Gap-sentences for Abstractive Summarization</article-title>
          ,
          <source>in: Proceedings of the 37th International Conference on Machine Learning (ICML)</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>11328</fpage>
          -
          <lpage>11339</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ghazvininejad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mohamed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          , L. Zettlemoyer, BART:
          <article-title>Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension</article-title>
          , arXiv preprint arXiv:
          <year>1910</year>
          .
          <volume>13461</volume>
          (
          <year>2019</year>
          ). URL: https://arxiv.org/abs/
          <year>1910</year>
          .13461.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <surname>M. La Quatra</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Caselle</surname>
          </string-name>
          , T. Oberhauser,
          <article-title>THExt: A Transformer-based Highlights Extractor for Scientific Papers</article-title>
          ,
          <source>in: Proceedings of the 29th International Conference on Computational Linguistics (COLING)</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>4932</fpage>
          -
          <lpage>4944</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2022</year>
          .coling-
          <volume>1</volume>
          .
          <fpage>435</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>A.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Patel</surname>
          </string-name>
          ,
          <article-title>Prompting Large Language Models for Scientific Highlight Generation</article-title>
          ,
          <source>arXiv preprint arXiv:2311.04567</source>
          (
          <year>2023</year>
          )
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>