<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Knowledge Extraction: Pegasus-driven Summarization for Research Dissemination</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Asha Hegde</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sharal Coelho</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Joiesmary D'Souza</string-name>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, Mangalore University</institution>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Manipal Academy of Higher Education</institution>
          ,
          <addr-line>Manipal</addr-line>
          ,
          <country>India, India</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2026</year>
      </pub-date>
      <abstract>
        <p>Many research articles in recent years have included research highlights to concisely summarize important ifndings, making it easier for academics to understand a paper's contribution and increasing the article's visibility through search engines. The field of automatic text processing research using Artificial Intelligence is very active recently. The proposed work uses a refined Pegasus model with Low-Rank Adaptation (LoRA) to summarize scientific abstracts into highlights for efective and high-quality summary generation. The process involves pre-processing and tokenizing the MixSub-SciHigh dataset provided by "Research Highlight Generation from Scientific Papers (SciHigh)" shared task, training with model variants from ArXiv and PubMed to provide accurate highlights from various scientific domain. Among the two models, the proposed summarization model that uses PubMed with LoRA obtained a ROUGE-L F1-score of 0.2296 and secured 4th rank in the shared task.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Text Summarization</kwd>
        <kwd>MixSub</kwd>
        <kwd>Research Highlight Generation</kwd>
        <kwd>Pegasus/PubMed</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The scientific research process begins with investigating the state of the art, which may involve a vast
number of publications [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. However, the growth in the number of scientific papers is generally viewed
as a positive development, but it also shows some challenges. With the exponential growth in the
number of scientific papers being published, it can be challenging for researchers to stay up to date
with the latest findings in their field and identify which papers are most important.
      </p>
      <p>
        Summarization is the process of reducing a source text into a shorter version while preserving
its information content [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Automatically summarizing scientific articles helps researchers in their
investigation by speeding up the research process [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. However, generating summaries of scientific
articles addresses the problems, such as (i) identifying the keywords that describe the main topics
covered by an article, (ii) generating an abstract of an article, or (iii) selecting the content that is most
likely to appear in the article highlights. Keywords are single words or phrases, which are usually
extracted using keyphrase extraction techniques. Abstracts are collections of whole sentences that
provide summaries of the most important information in a publication. Although they are commonly
available for most of the published articles, they can also be generated using extractive summarization
methods [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        Research highlights are straightforward, structured summaries that encapsulate the essential
contributions of a scientific paper, ofering a quick overview for students, journals, and researchers. Manually
preparing highlights is time-consuming, prompting the need for automated solutions. The shared task
Research Highlight Generation from Scientific Papers (SciHigh) focuses on automatically generating
research highlights from scientific paper abstracts. This shared task aims to develop Machine Learning
(ML) models to generate high-quality, author-like research highlights from abstracts using the MixSub
dataset [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], a subset of the MixSub corpus comprising 19,785 scientific papers from 2020. We explore
transformer-based models, retrieval-augmented approaches, and fine-tuned neural networks to achieve
this goal, addressing challenges such as incorrect information and factual inconsistencies. The task
evaluates submissions using ROUGE-1, ROUGE-2, ROUGE-L, and METEOR metrics, with rankings
based on the ROUGE-L F1-score. Our work aims to improve the eficiency and accuracy of highlight
generation, benefiting academic workflows and knowledge dissemination.
      </p>
      <p>The rest of the paper is organized as follows: Section 2 contains Related Work. While Section 3
describes the Methodology, Section 4 gives a description of the Datasets, Experiments, Results, and
Observations followed by Conclusion in Section 5.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Automatic text summarization has been extensively studied[
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], with significant advancements in
abstractive summarization driven by transformer-based models [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ][
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]. Models like BART [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], T5 [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ],
and PEGASUS [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] have shown strong performance in generating summaries for general and scientific
texts. While both BART and PEGASUS are designed for abstractive tasks, PEGASUS is better suited
for highlight-like outputs because it has been pre-trained on gap-sentence generation. Recent work
on scientific summarization, such as SciBERT [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] and PubMedBERT [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ], Utilizes domain-specific
pre-training to efectively manage specialized terminology, which is critical for the MixSub-SciHigh
dataset’s diverse scientific domains.
      </p>
      <p>
        Retrieval Augmented Generation (RAG) [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] combines retrieval and generation to improve accurate
consistency, as seen in studies like Multi-XScience [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] and SciTLDR[
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]. RAG retrieves relevant
context to ground outputs, addressing issues like hallucinations. Fine-tuning neural networks on
domain-specific datasets has also proven efective, with studies demonstrating improved ROUGE scores
through targeted fine-tuning on scientific texts. However, generating structured highlights remains
underexplored, and challenges like factual inconsistencies persist in scientific summarization. Nallapati
et al. [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] proposed an abstractive text summarization technique that uses attentional encoder-decoder
Recurrent Neural Networks (RNN). This model is used to generate a summary of a given input document.
Using a bidirectional RNN, the model first encodes the input document, which captures the input’s
contextual information. At the decoder end, the summary is then generated one word at a time by
considering the encoded input document. Our study builds on these advancements by focusing on
highlight generation, a summarization task, using the MixSub-SciHigh dataset. We aim to combine
transformer-based models, retrieval-augmented approaches, and fine-tuning strategies to optimize for
ROUGE-L F1-score.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>The objective of this work to generate research highlights using the abstraction of the research articles.
The methodology involves fine-tuning a pre-trained Pegasus model using Low-Rank Adaptation (LoRA)
for abstractive summarization of scientific articles, where abstracts are summarized into highlights.
The process begins with preprocessing steps followed by model building. Initially, the Pegasus-PubMed
model is used, and subsequently, it is replaced with the Pegasus-ArXiv model to compare performance
on domain-specific summarization tasks.</p>
      <sec id="sec-3-1">
        <title>3.1. Pre-processing</title>
        <p>The dataset consists of columns ’Abstract’ and ’Highlights’ which represent the input and target values,
respectively. The Pegasus tokenizer is applied to both inputs and targets, truncating to maximum
lengths of 256 and 64 tokens, respectively, with padding to ensure uniform sequence lengths. This
tokenized data is then mapped across the datasets, removing original columns to create model-ready
inputs.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Text Representation Using Pegasus/PubMed</title>
        <p>
          Pegasus1 is a pre-trained transformer-based model designed specifically for abstractive text
summarization [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ]. It uses a sequence-to-sequence architecture where the encoder processes the input text
and the decoder generates the summary, usig self-supervised pre-training on large corpora with a
gap-sentence generation objective to mimic summarization tasks. Its pre-training focuses on removing
important sentences from documents and generating them as pseudo-summaries, which improves
its ability to capture important points. The significance of the PubMed variant lies in its fine-tuning
on the PubMed dataset, a vast collection of biomedical literature abstracts, which provides it with
domain-specific knowledge in medical and life sciences terminology.
        </p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Text Representation Using Pegasus/ArXiv</title>
        <p>
          Pegasus is used due to its eficiency in abstractive summarization tasks. The ArXiv variant is fine-tuned
on the ArXiv dataset, which comprises millions of preprints in fields like physics, mathematics, computer
science, etc., allowing the model to specialize in non-biomedical scientific literature.
3.4. LoRA
LoRA[
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] is a Parameter-eficient Fine-tuning (PEFT) method that maintains competitive performance
while drastically reducing memory and computing expenses. In summarization tasks, LoRA’s
significance lies in its ability to fine-tune large models like Pegasus on domain-specific data with limited
computational resources. This is particularly beneficial for scientific summarization, where datasets
may be smaller or specialized, as LoRA preserves the model’s generalization while adapting to nuances
1https://huggingface.co/docs/transformers/model_doc/pegasus
like technical terminology. Furthermore, it facilitates experimentation, such as switching between
PubMed and ArXiv variants, by making the process more eficient and scalable. Overall, LoRA helps in
reducing memory footprint and training time without sacrificing summary quality.
        </p>
      </sec>
      <sec id="sec-3-4">
        <title>3.5. Model Construction</title>
        <p>The model is constructed by loading the pre-trained Pegasus base from either the PubMed or ArXiv
checkpoint. LoRA configuration is applied with r=8, alpha =16, and dropout =0.1. A Seq2SeqTrainer
is initialized with the adapted model. The training arguments includes batch sizes, learning rate, and
epochs, along with a data collator.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments an Result</title>
      <sec id="sec-4-1">
        <title>4.1. Dataset</title>
        <p>The proposed work utilizes a subset of the MixSub dataset. The MixSub corpus is created by collecting
research articles from ScienceDirect2, encompassing a diverse range of scientific domains. It contains
19,785 research papers published in the year 2020. The MixSub-SciHigh dataset is split into three sets
and detailed distribution is shown in table 1. Each data instance is structured as a pair consisting of the
abstract and the corresponding author-written research highlights. Each entry in the dataset includes:
• Abstract: A concise summary of the research paper.</p>
        <p>• Research Highlights: Key contributions manually written by the authors.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Experimental Setup</title>
        <p>The experiments are performed using the MixSub-SciHigh dataset from the "Research Highlight
Generation from Scientific Papers (SciHigh)" shared task. The dataset comprises scientific abstracts from
diverse fields, pre-processed and tokenized for model training. Two model variants are evaluated:
ifne-tuned on (i) ArXiv data and (ii) PubMed data, both utilizing the Pegasus model with LoRA for
eficient abstract summarization. The performance of the models is evaluated using ROUGE metrics
such as ROUGE-1, ROUGE-2, and ROUGE-L to measure the quality and relevance of the generated
research highlights.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Results</title>
        <p>The performance of the two model variants is summarized in the Table 2, based on ROUGE F1-scores
for the generated highlights. Figure 2 presents a bar chart comparing our team’s (NLPFusion) best
submission against other teams’ submissions in the shared task, highlighting our competitive performance
in ranking quality.</p>
        <p>The performance of our 2 submission runs for the SciHigh task is presented in Table 2. The
PubMedbased model outperformed the ArXiv-based model across all ROUGE metrics, achieving a ROUGE-L
F1-score of 0.2296, which secured the 4th rank in the SciHigh shared task. The higher performance of the
PubMed model can be attributed to its fine-tuning on biomedical literature, which aligns closely with the
domain-specific characteristics of the MixSub-SciHigh dataset. The results demonstrate the efectiveness
of the LoRA-based fine-tuning approach in generating accurate research highlights, contributing to
eficient summarization of scientific abstracts.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>The task of automatically generating research highlights from abstracts represents an important step
toward facilitating academic workflows. By using transformer-based models, our team, NLPFusion,
aims to produce high-quality, author-like highlights that achieve strong ROUGE-L F1-scores while
addressing challenges like hallucinations and factual inconsistencies. The given dataset’s diverse
scientific domains and structured highlight format provide a robust research facility for advancing
summarization techniques. Our work has the potential to reduce the burden on researchers and enhance
the accessibility of scientific contributions through automated, accurate highlight generation.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>In preparing this work, the author(s) utilized Grok3 for grammar and spelling checks. Paraphrasing
was handled via QuillBot. With this tool, the author(s) reviewed and revised the content as required,
while assuming full responsibility for the publication’s integrity.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>T.</given-names>
            <surname>Rehman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. K.</given-names>
            <surname>Sanyal</surname>
          </string-name>
          , S. Chattopadhyay,
          <article-title>Research highlight generation with elmo contextual embeddings</article-title>
          ,
          <source>Scalable Computing: Practice and Experience</source>
          <volume>24</volume>
          (
          <year>2023</year>
          )
          <fpage>181</fpage>
          -
          <lpage>190</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>K.</given-names>
            <surname>Woodsend</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lapata</surname>
          </string-name>
          ,
          <article-title>Automatic generation of story highlights, in: Proceedings of the 48th Annual Meeting of the Association for Computational Linguistics</article-title>
          , Association for Computational Linguistics,
          <year>2010</year>
          , pp.
          <fpage>565</fpage>
          -
          <lpage>574</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>T.</given-names>
            <surname>Rehman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Das</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. K.</given-names>
            <surname>Sanyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chattopadhyay</surname>
          </string-name>
          ,
          <article-title>An analysis of abstractive text summarization using pre-trained models</article-title>
          ,
          <source>in: Proceedings of International Conference on Computational Intelligence, Data Science and Cloud Computing: IEM-ICDC 2021</source>
          , Springer,
          <year>2022</year>
          , pp.
          <fpage>253</fpage>
          -
          <lpage>264</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S.</given-names>
            <surname>Chopra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Auli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Rush</surname>
          </string-name>
          ,
          <article-title>Abstractive sentence summarization with attentive recurrent neural networks, in: Proceedings of the 2016 conference of the North American chapter of the association for computational linguistics: human language technologies</article-title>
          ,
          <year>2016</year>
          , pp.
          <fpage>93</fpage>
          -
          <lpage>98</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>T.</given-names>
            <surname>Rehman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. K.</given-names>
            <surname>Sanyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chattopadhyay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Bhowmick</surname>
          </string-name>
          ,
          <string-name>
            <surname>P. P. Das</surname>
          </string-name>
          ,
          <article-title>Generation of highlights from research papers using pointer-generator networks and scibert embeddings</article-title>
          ,
          <source>IEEE Access 11</source>
          (
          <year>2023</year>
          )
          <fpage>91358</fpage>
          -
          <lpage>91374</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>T.</given-names>
            <surname>Rehman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Chattopadhyay</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. K.</given-names>
            <surname>Sanyal</surname>
          </string-name>
          ,
          <article-title>Abstractive summarization of scientific documents: Models and evaluation techniques</article-title>
          ,
          <source>in: Proceedings of the 15th Annual Meeting of the Forum for Information Retrieval Evaluation</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>121</fpage>
          -
          <lpage>124</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A. S.</given-names>
            <surname>Bashir</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Bichi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>U.</given-names>
            <surname>Mahmud</surname>
          </string-name>
          ,
          <string-name>
            <surname>A. M.</surname>
          </string-name>
          <article-title>Bello, Long-text abstractive summarization using transformer models: A systematic review</article-title>
          ,
          <source>Journal of the Brazilian Computer Society</source>
          <volume>31</volume>
          (
          <year>2025</year>
          )
          <fpage>1264</fpage>
          -
          <lpage>1279</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Jons</surname>
          </string-name>
          ,
          <article-title>Text summarization using a transformer architecture: An attention based transformer approach to abstractive summarization</article-title>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ghazvininejad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mohamed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Levy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Stoyanov</surname>
          </string-name>
          , L. Zettlemoyer, Bart:
          <article-title>Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension</article-title>
          , arXiv preprint arXiv:
          <year>1910</year>
          .
          <volume>13461</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>C.</given-names>
            <surname>Rafel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Shazeer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Roberts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Narang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Matena</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. J.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <article-title>Exploring the limits of transfer learning with a unified text-to-text transformer</article-title>
          ,
          <source>Journal of machine learning research 21</source>
          (
          <year>2020</year>
          )
          <fpage>1</fpage>
          -
          <lpage>67</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Saleh</surname>
          </string-name>
          , P. Liu, Pegasus:
          <article-title>Pre-training with extracted gap-sentences for abstractive summarization</article-title>
          ,
          <source>in: International conference on machine learning, PMLR</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>11328</fpage>
          -
          <lpage>11339</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>I.</given-names>
            <surname>Beltagy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Cohan</surname>
          </string-name>
          ,
          <article-title>Scibert: A pretrained language model for scientific text</article-title>
          , arXiv preprint arXiv:
          <year>1903</year>
          .
          <volume>10676</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Tinn</surname>
          </string-name>
          , H. Cheng, M. Lucas,
          <string-name>
            <given-names>N.</given-names>
            <surname>Usuyama</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Naumann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Poon</surname>
          </string-name>
          ,
          <article-title>Domainspecific language model pretraining for biomedical natural language processing</article-title>
          ,
          <source>ACM Transactions on Computing for Healthcare (HEALTH) 3</source>
          (
          <issue>2021</issue>
          )
          <fpage>1</fpage>
          -
          <lpage>23</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xiong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Jia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Retrieval-augmented generation for large language models: A survey</article-title>
          ,
          <source>arXiv preprint arXiv:2312.10997 2</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Lu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Dong</surname>
          </string-name>
          ,
          <string-name>
            <surname>L.</surname>
          </string-name>
          <article-title>Charlin, Multi-xscience: A large-scale dataset for extreme multi-document summarization of scientific articles</article-title>
          , arXiv preprint arXiv:
          <year>2010</year>
          .
          <volume>14235</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>I.</given-names>
            <surname>Cachola</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Cohan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. S.</given-names>
            <surname>Weld</surname>
          </string-name>
          ,
          <article-title>Tldr: Extreme summarization of scientific documents</article-title>
          , arXiv preprint arXiv:
          <year>2004</year>
          .
          <volume>15011</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>R.</given-names>
            <surname>Nallapati</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Gulcehre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Xiang</surname>
          </string-name>
          , et al.,
          <article-title>Abstractive text summarization using sequenceto-sequence rnns and beyond</article-title>
          ,
          <source>arXiv preprint arXiv:1602.06023</source>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>N. V. K. S.</given-names>
            <surname>Dasari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sungheetha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G. L. V.</given-names>
            <surname>Mahesh</surname>
          </string-name>
          , G. Danesh,
          <string-name>
            <given-names>T.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <article-title>Text summarization using pegasus transformer model in machine learning</article-title>
          ,
          <source>in: 2025 3rd International Conference on Inventive Computing and Informatics (ICICI)</source>
          , IEEE,
          <year>2025</year>
          , pp.
          <fpage>786</fpage>
          -
          <lpage>788</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>J. Y.-C. Hu</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Su</surname>
            ,
            <given-names>E.-J.</given-names>
          </string-name>
          <string-name>
            <surname>Kuo</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          <string-name>
            <surname>Song</surname>
          </string-name>
          , H. Liu,
          <article-title>Computational limits of low-rank adaptation (lora) for transformer-based models</article-title>
          ,
          <source>arXiv preprint arXiv:2406.03136</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>