<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Code-Mixing and Script-Mixing in Indian Language Sum marization with Transformer Models</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Pulkit Chatwal</string-name>
          <email>Pulkitchatwal@gmail.com</email>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Amit Agarwal</string-name>
          <email>aagarwal3@cs.iitr.ac.in</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ankush Mittal</string-name>
          <email>dr.ankush.mittal@gmail.com</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Indian Languages, Text Summarization, Pre­Trained Model, Sequence­to­Sequence models, Multilingual Text</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>AICoE Wells Fargo International Solutions Private Limited</institution>
          ,
          <addr-line>Bangalore</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>COER University</institution>
          ,
          <addr-line>Roorkee</addr-line>
          <country country="IN">India</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Rajiv Gandhi Institute of Petroleum Technology</institution>
          ,
          <addr-line>Jais</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Summarization</institution>
          ,
          <addr-line>Transformer Models, Fine­Tuning, LLM</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>With the growing need for accessible information across diverse linguistic backgrounds, text summarization in multilingual contexts has become increasingly essential. Text summarization is a crucial task in natural language processing, particularly in multilingual settings involving Indian languages. This paper presents our approach for the FIRE 2024 task, where we leverage large transformer­based language models for summarizing Indian languages, addressing the linguistic diversity and frequent instances of code­mixing and script­mixing unique to this context. Our methodology incorporates both extractive and abstractive summarization techniques, optimized for Indian languages through advanced fine­tuning of models like mT5, IndicBART, and BART. While prompt engineering has predominantly been applied to English tasks, we adapt it alongside fine­tuning to enhance summarization performance and computational efficiency. Our models achieved top results in five languages-Hindi, Gujarati, English, Tamil, and Bengali-and ranked second in Telugu. These results demonstrate substantial improvements in summarization accuracy, underscoring our approach's efficacy in handling the complexities of Indian languages and advancing text processing in multilingual, mixed­language environments.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>With the vast amount of information generated daily across multiple languages, the ability to automat­
ically summarize text has become essential for efficient information consumption and accessibility. In
multilingual and multicultural societies like India, where linguistic diversity includes hundreds of lan­
guages and dialects, automatic summarization solutions are crucial for bridging communication gaps and
ensuring equitable access to information. However, these challenges are further intensified by the preva­
lence of code­mixing (the blending of two or more languages within a single text) and script­mixing (the
use of multiple writing systems), making conventional summarization methods insufficient. Addressing
these complexities is essential for developing summarization tools that serve diverse user groups and
linguistic contexts effectively.</p>
      <p>
        Traditional summarization approaches often fall short in managing these complexities, particularly in
multilingual settings. For instance, [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] highlights the limitations of conventional summarization tech­
niques when dealing with code­mixed text. They emphasize the need for advanced, automatic summa­
rization methods capable of processing complex, multi­modal data, such as text, images, and audio, to
meet strategic information needs. Similarly, [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] explore the challenges faced by multilingual users who
frequently engage in code­mixing and underscore the necessity for conversational agents designed to
process mixed­language content effectively.
      </p>
      <p>To address these challenges, we propose a dual approach that leverages both fine­tuning and prompt­
based techniques applied to transformer­based models. Our solution involves fine­tuning multilingual
models (mT5 and IndicBART) and English models (BART) on a diverse dataset of Indian languages to
capture unique linguistic patterns, including code­mixing and script­mixing. Additionally, we employ</p>
      <p>CEUR</p>
      <p>ceur-ws.org
prompt engineering techniques to optimize summarization performance and computational efficiency,
particularly for English­language content. This combined approach creates a robust, adaptable solution
capable of generating accurate, concise summaries across a spectrum of Indian languages and mixed­
language inputs.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        Recent advancements in text summarization have employed a range of methods and datasets, particu­
larly focusing on fine­tuning transformer­based models for improved summarization performance. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]
introduced WikiLingua, a multilingual dataset with article­summary pairs in 18 languages, and fine­
tuned the mBART model on this dataset. While effective, these efforts mainly focused on language pairs
and did not address the complexities associated with code­mixing and script­mixing, both of which are
crucial for multilingual contexts such as Indian languages. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] utilized transformer­based models like
RoBERTa­Base and Flan T5 Base for cross­platform age classification on social media, achieving im­
pressive accuracy. [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] explored leadership traits during natural hazards by analyzing personality and
emotional characteristics, uncovering key differences between local and global leaders. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] introduced
AgriLLM, leveraging transformers to automate query resolution for farmers and bridge information gaps
in agriculture. [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] developed an SMS­based FAQ retrieval system using machine learning to refine noisy
text and improve information access. Similarly, [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] fine­tuned T5, BART, and Pegasus for abstractive
summarization of medical documents using the SUMPUBMED dataset, but their research did not extend
to the multilingual, mixed­language contexts typical of Indian languages.
      </p>
      <p>
        In the domain of extractive summarization, [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] proposed a graph­based approach that transformed
text into a network of interconnected sentences and used a selectivity measure to assess node signifi­
cance. This graph­based model, while innovative, is limited in handling language­specific nuances and
lacks adaptability for abstractive summarization in mixed­language texts. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] provided a comparative
analysis of major extractive summarization techniques such as TF­IDF, Clustering, Fuzzy Logic, Neu­
ral Networks, and Graph­based methods, yet these methods typically struggle with capturing abstracted
meaning, particularly in mixed­language settings.
      </p>
      <p>
        For English text summarization, [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] demonstrated the effectiveness of BERT­based models across var­
ious datasets, showcasing strong summarization quality. Similarly, [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] explored reinforcement learning
for abstractive summarization, optimizing both readability and content fidelity. However, these methods
were focused on monolingual English text and did not address the unique challenges of Indian languages,
where code­mixing and script­mixing are common.
      </p>
      <p>
        In language­specific research, [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] used TextRank and Fuzzy C­means for Bengali text summariza­
tion, highlighting the importance of customized models for individual languages. For Gujarati, [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]
combined pre­trained language models with clustering techniques, showing promising results for low­
resource languages. However, these methods were largely extractive and were limited in their ability to
handle code­mixed content across multiple Indian languages.
      </p>
      <p>
        Studies from shared tasks organized by FIRE ([
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]; [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ]; [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]; [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]; [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]; [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]) have examined sum­
marization challenges in Indian languages, employing diverse methodologies and models. Although
these contributions have advanced summarization for Indian languages, they primarily focused on single­
language settings or basic multilingual scenarios, lacking robust solutions for complex, code­mixed in­
puts.
      </p>
      <p>Our Approach: Our work differs from previous studies by specifically targeting the multilingual,
code­mixed, and script­mixed text summarization challenges inherent in Indian languages. We adopt a
dual approach that combines fine­tuning and prompt­based methods for transformer­based models, such
as mT5, IndicBART, and BART. By fine­tuning these models on a dataset of Indian languages and lever­
aging prompt engineering for computational efficiency, our approach is designed to capture the unique
linguistic patterns of each language, including mixed­language constructs. This combined methodology
allows us to generate coherent, high­quality summaries that address the specific complexities of Indian
language contexts, setting our work apart as a comprehensive solution for multilingual summarization</p>
    </sec>
    <sec id="sec-3">
      <title>3. Problem Statement</title>
      <p>Let  represent the dataset of news articles across multiple Indian languages, where:
 = {(  ,   ) ∣   ∈ ,   ∈ } =1
where: ­   is the  ­th article in the dataset, containing a mixture of text from multiple languages and
potential code­mixing and script­mixing. ­   is the corresponding reference summary for   . ­  is
the space of input texts (articles), and  is the space of target summaries. ­  is the total number of
article­summary pairs in the dataset.</p>
      <p>The goal is to learn a mapping function  ∶  → 
that generates concise and informative summaries
for each article   such that the generated summary  ̂ =  (  ) approximates   .</p>
      <sec id="sec-3-1">
        <title>3.1. Problem Objective</title>
        <p>The objective is to minimize the error between the generated summaries  ̂ and the reference summaries   ,
typically measured using evaluation metrics such as ROUGE, BLEU, or cosine similarity in embedding
space. Formally:</p>
        <p>min ∑ ℒ ( (
 =1
 ),   )
where: ­ ℒ is the loss function representing the error between the generated summary  ̂ =  (  ) and
the reference summary   .
3.2. Additional Constraints and Considerations
handle cross­lingual transfer effectively.</p>
        <p>| ̂ | ≈  , where  is the desired summary length.
• Multilingual and Code­Mixed Text:   may contain tokens from multiple languages  =
{ 1,  2, … ,   }, where each   corresponds to a distinct language script. Thus, the model  must
• Length Constraint: Each generated summary  ̂ should ideally satisfy a fixed length constraint
• Semantic Fidelity: The mapping  should retain the essential semantic information from   in  ̂ ,
aligning with the reference summaries   in terms of main facts and insights.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Model Description</title>
      <p>
        English Language Model: BART For English­language summarization, we employ the facebook/bart­
large­cnn model, a state­of­the­art transformer­based encoder­decoder (seq2seq) architecture. BART
combines a bidirectional encoder, akin to BERT, with an autoregressive decoder, similar to GPT. This
model is pre­trained with a denoising autoencoder objective, where it learns to reconstruct text corrupted
by noise functions. This pre­training process equips BART to handle various downstream tasks, includ­
ing summarization and translation, as well as comprehension tasks like text classification and question
answering. For our experiments, we fine­tuned the model on the CNN/Daily Mail dataset, which con­
tains a substantial collection of paired text and summary samples, ensuring robust performance in text
summarization tasks [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ].
      </p>
      <p>Multilingual Language Models: mT5 and IndicBART To address summarization for multiple
Indian languages, such as Gujarati, Telugu, and Bengali, we leverage the csebuetnlp/mT5_multilin­
gual_XLSum model. mT5 is a multilingual variant of T5, designed to handle diverse languages by using a
shared vocabulary and multilingual training data. For this task, we utilize XL­Sum, a high­quality, mul­
tilingual dataset curated for abstractive summarization with approximately 1 million article­summary
pairs sourced from BBC News across 44 languages, including low­resource languages. XL­Sum empha­
sizes abstractive summarization with a high level of brevity, abstraction, and quality, as indicated by both
human judgments and intrinsic metrics [22].</p>
      <p>Hindi and Tamil Language Model: IndicBART For Hindi and Tamil summarization tasks, we em­
ploy ai4bharat/IndicBART, a multilingual sequence­to­sequence model focused on Indian languages. In­
dicBART, based on the mBART architecture, is specifically tailored for 11 Indian languages and supports
natural language generation tasks like summarization and machine translation. This model has been pre­
trained on an extensive corpus of Indic languages, containing 452 million sentences and 9 billion tokens,
where all languages are transcribed into the Devanagari script to facilitate cross­lingual transfer learning.
This approach enhances its performance in resource­constrained Indic languages by effectively leverag­
ing syntactic and semantic similarities across languages [23].</p>
    </sec>
    <sec id="sec-5">
      <title>5. Dataset Description</title>
      <p>The dataset assigned to the ILSUM 2024 task is comprehensive and extends the groundwork set by pre­
vious editions, and support for three more Dravidian languages is added: Kannada, Tamil, and Telugu.
Added datasets enhance the coverage of regional Indian languages in the text summarization space and
continue the trend in previous years [24] . Each dataset is collected from main newspapers and arranged
to support both extractive and abstractive summarization methodologies. The number of document­
summary pairs may well be a good basis for model formulation and evaluation related to this task.</p>
      <p>A characteristic of this year's dataset is the prevalence of code­mixing and script­mixing, which poses
a unique challenge to the language models. Code­mixing here refers to the use of English phrases within
articles that are essentially composed in Indian languages, a common occurrence within the country's
media environment. This happens quite frequently in headlines and news stories, making it a significant
challenge for summarization models. For example:</p>
      <p>Example of code­mixing in a news article:
• IND vs SA, T20 તસવીરોમાં : વરસાદે વલન બની મજા બગાડી! (India vs SA, 5th T20 in pictures: rain
spoils the match)
• LIC IPO में पैसा लगाने वालाें का टूटा दल, आई एक और मुक़सानदे ह खबर (Investors of LIC IPO left broken­
hearted, yet another bad news)
• Hubballi Special Trains: ಹుుಿಯಂದ ದೆಹಲಿ ಈ ನಗರ ಿವ ೕ ಷ ೖ ಲು ಆರಂಭ (Special train starts from</p>
      <p>Hubballi to this city of the country)</p>
      <p>The dataset is divided into separate CSV files for each language, which are Hindi, Gujarati, Bengali,
Tamil, Telugu, Kannada, and English. Each file contains columns that represent the source text and
its corresponding summary, which gives a strong foundation for training and testing the models. The
integration of the three Dravidian languages is a big leap in this year's work, indicating an ongoing effort
to increase the diversity of the language representation in the Indian language summarization task [25].</p>
    </sec>
    <sec id="sec-6">
      <title>6. Method</title>
      <sec id="sec-6-1">
        <title>6.1. Task Description</title>
        <p>The task involves generating concise, informative, and fixed­length summaries for news articles in mul­
tiple Indian languages, addressing the complexities of code­mixing and script­mixing, where languages
often blend within the same text. Our dataset comprises headline­article pairs sourced from major news­
papers in languages including Tamil, Gujarati, Telugu, Bengali, and Kannada. This multilingual dataset
introduces diverse linguistic structures, making it ideal for evaluating and refining summarization capa­
bilities across mixed­language content.</p>
      </sec>
      <sec id="sec-6-2">
        <title>6.2. Core Methodology</title>
        <p>To address the challenges of multilingual summarization, we implemented a dual approach involving
model fine­tuning and prompt engineering. This methodology facilitated efficient handling of diverse
linguistic inputs while preserving the quality of generated summaries.</p>
        <p>Model Fine­Tuning: mT5, IndicBART, and BART</p>
        <p>We fine­tuned three pre­trained models—mT5, IndicBART, and BART—on our dataset containing
thousands of document­summary pairs across various languages. Fine­tuning enabled the models to
adapt to specific linguistic features, handling both code­mixing and script­mixing effectively.
• mT5: Targeted for Gujarati, Telugu, and Bengali summarization, mT5 leverages its multilingual ar­
chitecture to support cross­lingual summarization. By fine­tuning mT5 on our dataset, we utilized
its cross­lingual transfer capabilities, which enhanced performance in lower­resource settings, par­
ticularly for languages like Tamil and Telugu, where English words often appear within the native
language text.
• IndicBART: Applied for Hindi and Tamil, IndicBART, designed specifically for text generation
tasks in Indic languages, demonstrated computational efficiency and strong summarization perfor­
mance. Fine­tuning this model on our dataset allowed it to handle code­mixing by leveraging its
foundational understanding of Indic language syntax and semantics.
• BART: For English summaries, we used BART, a transformer­based seq2seq model pre­trained
on news articles. Fine­tuning BART on our dataset optimized its capability to produce coherent
and compact summaries of English content, capturing complex information effectively.
Prompt Engineering for English Summarization</p>
        <p>Alongside fine­tuning, we utilized prompt engineering specifically for English summarization to re­
duce computational overhead. This approach uses task­specific prompts to guide BART’s summarization
capabilities without additional model retraining. By designing prompts tailored to the summarization
task, we achieved efficient, high­quality summaries with reduced resource demands.</p>
        <p>Example Prompts:
• "Summarize the following article clearly and concisely, emphasizing the main facts, insights, and
key points. Exclude extraneous details, aiming for a natural and human­like flow. Target summary
length: 45­90 words."
• "Create a semantically rich summary of the following article, ensuring coverage of core messages,
facts, and meaning. The summary should be concise yet comprehensive, maintaining accuracy and
coherence in a way that retains the essence of the original content."</p>
        <p>Comparison of Fine­Tuning and Prompt Engineering Approaches While fine­tuning produced
slightly higher accuracy, prompt engineering was highly efficient, particularly for English articles, reduc­
ing computational time and resource usage. By combining both approaches, we achieved an effective
balance between performance and computational efficiency.</p>
        <p>Generated Summary Examples:
Where:
Where:</p>
      </sec>
      <sec id="sec-6-3">
        <title>7.2. BERT Score</title>
        <p>value offers."
• Fine­Tuning: "Despite significant investments in star players like Cristiano Ronaldo, Neymar,
and Karim Benzema, the Saudi Pro League was unable to secure Lionel Messi, even after high­
• Prompt Engineering: "Lionel Messi expressed interest in joining Cristiano Ronaldo in the ‘pow­
erful’ Saudi Pro League before transferring to MLS. Following his departure from Paris Saint­
Germain, Messi joined Inter Miami on a free transfer. Recently, TIME magazine honored him as
‘Athlete of the Year.’"</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>7. Evaluation Metrics</title>
      <sec id="sec-7-1">
        <title>The summarization models are evaluated using ROUGE and BERT scores:</title>
        <sec id="sec-7-1-1">
          <title>7.1. ROUGE (Recall-Oriented Understudy for Gisting Evaluation)</title>
          <p>Measures n­gram overlap between generated and reference summaries, capturing how much key infor­
mation from the reference is present.</p>
          <p>ROUGE­N =
∑Countmatched( ­gram)</p>
          <p>∑Counttotal( ­gram)
• Count_matched: Number of matching n­grams between the generated and reference summaries.
• Count_total: Total number of n­grams in the reference summary.</p>
          <p>Compares the semantic similarity between generated and reference summaries using BERT embeddings.</p>
          <p>BERT Score = 1</p>
          <p>=1</p>
          <p>∑ max cosine_similarity(BERT(  ), BERT(  ))
•   : Token in the generated summary.
•   : Token in the reference summary.</p>
          <p>• Cosine similarity: Measures how similar the meaning of tokens is.</p>
        </sec>
        <sec id="sec-7-1-2">
          <title>7.3. BERT Precision &amp; Recall</title>
          <p>Precision: Measures how much the generated summary matches the reference.</p>
        </sec>
      </sec>
      <sec id="sec-7-2">
        <title>Where  represents the tokens in the generated summary.</title>
        <p>Recall: Measures how much of the reference is covered by the generated summary.</p>
        <p>Precision =</p>
        <p>∑ max cosine_similarity(BERT(  ), BERT(  ))
Recall =
∑ max cosine_similarity(BERT(  ), BERT(  ))
1
|| 
1
||</p>
        <p />
        <sec id="sec-7-2-1">
          <title>7.4. BERT F1 Score</title>
        </sec>
      </sec>
      <sec id="sec-7-3">
        <title>Balances precision and recall:</title>
        <p>1 = 2 ×</p>
      </sec>
      <sec id="sec-7-4">
        <title>Precision × Recall</title>
      </sec>
      <sec id="sec-7-5">
        <title>Precision + Recall</title>
        <p>This evaluation combines both n­gram overlap (ROUGE) and semantic understanding (BERT).</p>
      </sec>
    </sec>
    <sec id="sec-8">
      <title>8. Results</title>
      <p>The official results for the ILSUM 2024 [26] challenge demonstrate the strong performance of our team,
Data Lovers, in multilingual summarization across several Indian languages. We participated in the
Hindi, English, Tamil, Telugu, Bengali, and Gujarati subtracks of Task 1. Our model achieved first place
in Hindi, English, Tamil, Bengali, and Gujarati, while we ranked second in Telugu. These high ranks
underscore the robustness and adaptability of our approach in handling diverse languages with varied
linguistic features. The official ROUGE and BERT scores for each language are presented in Table 2
and Table 3.</p>
      <p>Table 2 highlights our ROUGE score performance across different languages. Our model achieved the
highest ROUGE­1 score in Hindi (0.3659) and English (0.3644), closely followed by Telugu (0.3022).
This suggests that our model effectively captures essential information and meaning across languages.
Notably, the performance in Gujarati, Bengali, and Tamil was comparatively lower but still competitive,
reflecting the effectiveness of our approach in low­resource languages as well. The high ROUGE­L
scores, particularly in Hindi (0.3388) and English (0.3133), indicate that our model maintained coherence
and fluency in its generated summaries, a crucial aspect in multilingual summarization tasks.</p>
      <p>Table 3 shows the BERT scores, which measure semantic similarity between the generated summaries
and reference summaries. The English subtrack achieved the highest BERTScore­F1 (0.8781), show­
casing the model's superior ability to retain semantic meaning in English. For Indian languages, Telugu
(0.7532), Hindi (0.7396), and Gujarati (0.7398) displayed strong performance, indicating the model's ca­
pacity to handle linguistic nuances across Indian languages. The consistently high BERTScore­Precision
and Recall across languages reflect our model's reliability in generating summaries that closely match
the reference texts in meaning and structure.</p>
      <p>In the English subtrack, we conducted experiments comparing two methodologies: fine­tuning and
prompt engineering. While fine­tuning provided the highest accuracy in both ROUGE and BERT met­
rics, the prompt engineering approach was more computationally efficient, achieving a rank of 4th overall
in ROUGE and 5th in BERT scores. Table 4 and Table 5 present a comparative analysis between the two
methods. In Table 4, fine­tuning with BART achieved higher ROUGE­1 and ROUGE­L scores (0.3644
and 0.3133, respectively), highlighting its effectiveness in producing summaries with high lexical sim­
ilarity to the reference. Prompt engineering, despite lower ROUGE scores (ROUGE­1 of 0.3238 and
ROUGE­L of 0.2806), demonstrated considerable potential for applications requiring lower computa­
tional costs without substantial quality trade­offs.</p>
      <p>As shown in Table 5, fine­tuning outperformed prompt engineering in terms of BERTScore­F1 (0.8781
compared to 0.8687), indicating its superior ability to maintain semantic fidelity. However, prompt engi­
neering achieved comparable BERTScore­Recall (0.8847), indicating that it captures essential informa­
tion well, albeit with slightly less precision.</p>
      <p>Overall, our approach effectively balances performance and computational efficiency, making it adapt­
able for diverse use cases. The results affirm that fine­tuning is highly effective for high­quality summa­
rization, while prompt engineering offers a viable alternative in resource­constrained settings.</p>
    </sec>
    <sec id="sec-9">
      <title>9. Conclusion &amp; Future Work</title>
      <p>In this study, we evaluated several large language models, including mT5, IndicBART, and BART, on
the task of generating fixed­length summaries of news articles in multiple Indian languages. Our results
demonstrate that finetuned models consistently outperformed other methods, achieving top ROUGE and
BERT metrics across five out of six languages. Specifically, finetuning yielded ROUGE­1 scores as high
as 0.3659 for Hindi and 0.3644 for English, while BERTScore­F1 reached 0.8781 for English, underscor­
ing the models’ robustness in handling the complexities of multilingual summarization, particularly with
challenges like codemixing and scriptmixing.</p>
      <p>This research also highlights the potential of prompt engineering, especially for English summariza­
tion, where it achieved competitive BERTScore­Recall (0.8847) and a notable 4th rank in ROUGE scores.
Although prompt engineering slightly underperformed compared to full finetuning, it reduced computa­
tional costs and processing times by approximately 30%, establishing it as a cost­effective alternative
for resource­constrained settings. However, the results also indicated that prompt engineering was less
effective for Indian languages, especially those with complex codemixed and scriptmixed text. Future
efforts should focus on finetuning prompt­based methods specifically for Indian languages to improve
performance in mixed­language contexts.</p>
      <p>Future work will explore larger, more advanced models like GPT­4, LLaMA 2, and BLOOM to further
improve accuracy and efficiency, particularly for low­resource languages. We aim to finetune these mod­
els for specific Indian languages and continue experimenting with prompt­based techniques to maximize
summarization quality. Additionally, we plan to develop refined prompt engineering strategies tailored to
codemixing and scriptmixing challenges in Indian languages. By combining prompt­driven approaches
with advanced models, we aim to build an efficient summarization framework that balances high quality
and computational efficiency, expanding accessibility and effectiveness across diverse linguistic settings.
10. Declaration on Generative AI</p>
      <sec id="sec-9-1">
        <title>No generative AI tools were used during the preparation of this work.</title>
        <p>[22] T. Hasan, A. Bhattacharjee, M. S. Islam, K. Samin, Y.­F. Li, Y.­B. Kang, M. S. Rahman, R. Shahri­
yar, Xl­sum: Large­scale multilingual abstractive summarization for 44 languages, arXiv preprint
arXiv:2106.13822 (2021).
[23] R. Dabre, H. Shrotriya, A. Kunchukuttan, R. Puduppully, M. M. Khapra, P. Kumar, Indicbart: A
pre­trained model for indic natural language generation, arXiv preprint arXiv:2109.02903 (2021).
[24] S. Satapara, P. Mehta, S. Modha, D. Ganguly, Indian language summarization at fire 2023, in:
Proceedings of the 15th Annual Meeting of the Forum for Information Retrieval Evaluation, 2023,
pp. 27­­29.
[25] S. Satapara, P. Mehta, S. Modha, A. Hegde, S. HL, D. Ganguly, Overview of the third shared task
on indian language summarization (ilsum 2024), in: K. Ghosh, T. Mandl, P. Majumder, D. Ganguly
(Eds.), Working Notes of FIRE 2024 ­ Forum for Information Retrieval Evaluation, Gandhinagar,
India. December 12­15, 2024, CEUR Workshop Proceedings, CEUR­WS.org, 2024.
[26] S. Satapara, P. Mehta, S. Modha, A. Hegde, S. HL, D. Ganguly, Key insights from the third ilsum
track at fire 2024, in: Proceedings of the 16th Annual Meeting of the Forum for Information
Retrieval Evaluation, FIRE 2024, Gandhiinagar, India. December 12­15, 2024, ACM, 2024.</p>
      </sec>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Jangra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Mukherjee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jatowt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Saha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hasanuzzaman</surname>
          </string-name>
          ,
          <article-title>A survey on multi­modal summarization</article-title>
          ,
          <source>ACM Computing Surveys</source>
          <volume>55</volume>
          (
          <year>2023</year>
          )
          <fpage>1</fpage>
          ­­
          <lpage>36</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Y. J.</given-names>
            <surname>Choi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <article-title>Toward a multilingual conversational agent: Challenges and expectations of code­mixing multilingual users</article-title>
          ,
          <source>in: Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>1</fpage>
          ­­
          <lpage>17</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>F.</given-names>
            <surname>Ladhak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Durmus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Cardie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>McKeown</surname>
          </string-name>
          ,
          <string-name>
            <surname>Wikilingua:</surname>
          </string-name>
          <article-title>A new benchmark dataset for crosslingual abstractive summarization</article-title>
          , arXiv preprint arXiv:
          <year>2010</year>
          .
          <volume>03093</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>T.</given-names>
            <surname>Sankar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Suraj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Reddy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Toshniwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          , Iitroorkee@ smm4h
          <year>2024</year>
          cross
          <article-title>­platform age detection in twitter and reddit using transformer­based model</article-title>
          ,
          <source>in: Proceedings of The 9th Social Media Mining for Health Research and Applications (SMM4H</source>
          <year>2024</year>
          )
          <article-title>Workshop</article-title>
          and Shared Tasks,
          <year>2024</year>
          , pp.
          <fpage>101</fpage>
          ­­
          <lpage>105</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>A.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Toshniwal</surname>
          </string-name>
          ,
          <article-title>Identifying leadership characteristics from social media data during natural hazards using personality traits</article-title>
          ,
          <source>Scientific reports 10</source>
          (
          <year>2020</year>
          )
          <fpage>2624</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>K.</given-names>
            <surname>Didwania</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Seth</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kasliwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          , Agrillm:
          <article-title>Harnessing transformers for farmer queries</article-title>
          ,
          <source>arXiv preprint arXiv:2407.04721</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Bhatt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mittal</surname>
          </string-name>
          ,
          <article-title>Construction of a semi­automated model for faq retrieval via short message service</article-title>
          ,
          <source>in: Proceedings of the 7th Annual Meeting of the Forum for Information Retrieval Evaluation</source>
          ,
          <year>2015</year>
          , pp.
          <fpage>35</fpage>
          ­­
          <lpage>38</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>E.</given-names>
            <surname>Lalitha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ramani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Shahida</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. V. S.</given-names>
            <surname>Deepak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. H.</given-names>
            <surname>Bindu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Shaikshavali</surname>
          </string-name>
          ,
          <article-title>Text summarization of medical documents using abstractive techniques</article-title>
          ,
          <source>in: 2023 2nd International Conference on Applied Artificial Intelligence and Computing (ICAAIC)</source>
          , IEEE,
          <year>2023</year>
          , pp.
          <fpage>939</fpage>
          ­­
          <lpage>943</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>S.</given-names>
            <surname>Gowhar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Gupta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Madasamy</surname>
          </string-name>
          ,
          <article-title>Advancing human­like summarization: Approaches to text summarization</article-title>
          .,
          <source>in: FIRE (Working Notes)</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>747</fpage>
          ­­
          <lpage>754</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>K.</given-names>
            <surname>Jewani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Damankar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Janyani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Mhatre</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gangwani</surname>
          </string-name>
          ,
          <article-title>A brief study on approaches for extractive summarization</article-title>
          ,
          <source>in: 2021 International Conference on Artificial Intelligence and Smart Systems (ICAIS)</source>
          , IEEE,
          <year>2021</year>
          , pp.
          <fpage>601</fpage>
          ­­
          <lpage>608</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A.</given-names>
            <surname>Agrawal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Jain</surname>
          </string-name>
          , Divanshi,
          <string-name>
            <given-names>K.</given-names>
            <surname>Seeja</surname>
          </string-name>
          ,
          <article-title>Text summarisation using bert</article-title>
          ,
          <source>in: International Conference On Innovative Computing And Communication</source>
          , Springer,
          <year>2023</year>
          , pp.
          <fpage>229</fpage>
          ­­
          <lpage>242</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>R.</given-names>
            <surname>Paulus</surname>
          </string-name>
          ,
          <article-title>A deep reinforced model for abstractive summarization</article-title>
          ,
          <source>arXiv preprint arXiv:1705.04304</source>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>A.</given-names>
            <surname>Rahman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. M.</given-names>
            <surname>Rafiq</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Saha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Rafian</surname>
          </string-name>
          ,
          <article-title>Bengali text summarization using TextRank, Fuzzy C­means and aggregated scoring techniques</article-title>
          ,
          <source>Ph.D. thesis</source>
          , BRAC University,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>K.</given-names>
            <surname>Kumari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Kumari</surname>
          </string-name>
          ,
          <article-title>An extractive approach for automated summarization of indian languages using clustering techniques</article-title>
          .,
          <source>in: FIRE (Working Notes)</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>418</fpage>
          ­­
          <lpage>423</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>S.</given-names>
            <surname>Satapara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Modha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Modha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mehta</surname>
          </string-name>
          ,
          <article-title>Fire 2022 ilsum track: Indian language summarization</article-title>
          ,
          <source>in: Proceedings of the 14th Annual Meeting of the Forum for Information Retrieval Evaluation</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>8</fpage>
          ­­
          <lpage>11</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>S.</given-names>
            <surname>Satapara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Modha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Modha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mehta</surname>
          </string-name>
          ,
          <article-title>Findings of the first shared task on indian language summarization (ilsum): Approaches challenges and the path ahead</article-title>
          .,
          <source>in: FIRE (Working Notes)</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>369</fpage>
          ­­
          <lpage>382</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>S.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Deepak</surname>
          </string-name>
          ,
          <article-title>Deep learning based abstractive summarization for english language</article-title>
          .,
          <source>in: FIRE (Working Notes)</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>383</fpage>
          ­­
          <lpage>392</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>A.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Naik</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. S.</given-names>
            <surname>Sonawane</surname>
          </string-name>
          ,
          <article-title>Abstractive text summarization for hindi language using indicbart</article-title>
          .,
          <source>in: FIRE (Working Notes)</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>409</fpage>
          ­­
          <lpage>417</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>A.</given-names>
            <surname>Urlana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Bhatt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Surange</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Shrivastava</surname>
          </string-name>
          ,
          <article-title>Indian language summarization using pretrained sequence­to­sequence models</article-title>
          ,
          <source>arXiv preprint arXiv:2303.14461</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>S.</given-names>
            <surname>Satapara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mehta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Modha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ganguly</surname>
          </string-name>
          ,
          <article-title>Key takeaways from the second shared task on indian language summarization</article-title>
          (ilsum
          <year>2023</year>
          )., in: FIRE (Working Notes),
          <year>2023</year>
          , pp.
          <fpage>724</fpage>
          ­­
          <lpage>733</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>M.</given-names>
            <surname>Lewis</surname>
          </string-name>
          ,
          <article-title>Bart: Denoising sequence­to­sequence pre­training for natural language generation, translation, and comprehension</article-title>
          , arXiv preprint arXiv:
          <year>1910</year>
          .
          <volume>13461</volume>
          (
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>