=Paper=
{{Paper
|id=Vol-3395/T6-1
|storemode=property
|title=Findings of the First Shared Task on Indian Language Summarization (ILSUM): Approaches Challenges and the Path Ahead
|pdfUrl=https://ceur-ws.org/Vol-3395/T6-1.pdf
|volume=Vol-3395
|authors=Shrey Satapara,Bhavan Modha,Sandip Modha,Parth Mehta
|dblpUrl=https://dblp.org/rec/conf/fire/SataparaMM022a
}}
==Findings of the First Shared Task on Indian Language Summarization (ILSUM): Approaches Challenges and the Path Ahead==
Findings of the First Shared Task on Indian Language Summarization (ILSUM): Approaches, Challenges and the Path Ahead Shrey Satapara1 , Bhavan Modha2 , Sandip Modha3 and Parth Mehta4 1 Indian Institute of Technology, Hyderabad, India 2 University Of Texas at Dallas, USA 3 LDRP-ITR, Gandhinagar, India 4 Parmonic, USA Abstract This paper provides an overview of the first edition of the shared task on Indian Language Summa- rization (ILSUM) organized at the 14th Forum for Information Retrieval Evaluation (FIRE 2022). The objective of this shared task was to create benchmark data for text summarization in Indian languages. This edition included three languages Hindi, Gujarati and Indian English. Indian English is an officially recognised dialect of English mainly used in the Indian subcontinent. The combined train and test datasets included more than 10000 article-summary pairs for each language which, to the best of our knowledge, is the largest publicly available summarization dataset for Indian languages. The task saw an enthusiastic response, with registrations from over 50 teams. A total of 13 teams submitted runs across the three languages out of which 10 teams submitted working notes. Standard ROUGE metrics were used as the evaluation metric. Indian English saw the most enthusiastic response with all 10 teams participating, followed by 6 teams submitting runs for Hindi with 5 teams for Gujarati. Keywords Automatic Text Summarization, Indian Languages, Headline Generation 1. Introduction Research in Natural Language Processing has been known to be an uneven playing field for a long time. There is a chasm between the progress in resource-rich languages like English, Spanish, Chinese, etc as opposed to more resource-constrained languages like Hindi, Gujarati, Arabic, Urdu, etc. Although with the latest developments in the last few years, especially with open source large language models[1] and initiatives like the Forum for Information Re- trieval Evaluation (FIRE)[2], this gap is slowly bridging. The progress however has been task- dependent. For instance tasks like hate speech detection[3, 4, 5, 6, 7], Sentiment analysis[8, 9], mixed script IR[10, 11], Indian legal document retrieval and summarization[12, 13, 14, 15, 16], Fake news detection[17, 18], authorship attribution[19, 20] to name a few, have made progress Forum for Information Retrieval Evaluation, December 9-13, 2022, India £ shreysatapara@gmail.com (S. Satapara); bhavanmodha@gmail.com (B. Modha); sjmodha@gmail.com (S. Modha); parth.mehta126@gmail.com (P. Mehta) Ȉ 0000-0001-6222-1288 (S. Satapara); 0000-0002-1138-5968 (B. Modha); 0000-0003-2427-2433 (S. Modha); 0000-0002-4509-1298 (P. Mehta) © 2022 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). CEUR Workshop Proceedings http://ceur-ws.org ISSN 1613-0073 CEUR Workshop Proceedings (CEUR-WS.org) in past few years with several large scale datasets and pre-trained models becoming publicly available. Automatic text summarization on the other hand is one of the sub-disciplines of NLP where research is still more skewed towards English[21, 22, 23] and other resource-rich languages, while the focus on other resource-poor languages is almost negligible[24]. Indian languages, despite having millions of speakers, have received surprisingly little atten- tion. While on one hand large-scale datasets with hundreds of thousands of documents exist for languages like English[25], Chinese[26], Spanish[27], etc., the datasets for any Indian lan- guage runs into at most a few dozen documents[28, 29, 30, 31, 32, 33]. Further most existing datasets are either not public or are too small to be useful. As a result, hardly any meaningful research has been possible in this area. Through this shared task, we aim to bridge the existing gap by creating reusable corpora for Indian Language Summarization. In the first edition, we cover two major Indian languages Hindi and Gujarati, which have over 350 million and over 50 million speakers respectively. Apart from this we also include Indian English, a widely recognized dialect that can be substantially different from English spoken elsewhere. We provided over 10,000 news articles accompanied by a title and headlines for each language. Table 1 presents the details of the ILSUM dataset. The task is to generate a meaningful summary, either extractive or abstractive, for each article. 2. Related Work The first serious attempt at creating a reusable dataset for automatic text summarization was perhaps made during the Document Understanding Conference (DUC)[34] in 2002. The dataset was a collection of news articles on 50 topics and four handwritten summaries for each article. This was followed up in subsequent years with new additions and new tasks. The DUC was later followed by the Text Analysis Conference (TAC)[35]. TAC ran for several years and, like DUC, produced several benchmark corpora. On the whole, the DUC and TAC datasets together have been by far the most popular datasets for evaluating text summarization. However, with the advent of deep learning and large language models, the DUC and TAC corpora became inadequate because of smaller corpus sizes. Since then the focus shifted to large-scale datasets that can be used for training deep neural networks. Often these datasets were built by collect- ing already available article summary pairs, for example from newspapers, rather than creating the summaries. One such very popular dataset is the CNN/Dailymail dataset[25]. The dataset consists of around 300K articles from CNN and Dailymail newspapers, and the headlines of the articles were used as a multi-sentence summary. This strategy was often reused for En- glish as well as other languages. For instance, one of the largest Chinese datasets (LCTCS)[26] and Spanish (DACSA)[27] also employs the same strategy. A similar approach is also used for domain-specific summarizationParikh et al.. Compared to these the Indian language datasets are rather limited in size. Here we cover some of the more noteworthy attempts at creating text summarization datasets for Indian lan- guages. An exhaustive list of the datasets is available made available in [24]. The most popular and cited corpus is a Malayalam dataset that was developed using news articles and human- written summary pairs[33]. The corpus has 100 documents and is not released publicly. It is mainly used by the same research group for experimentation and there are no reports from other groups that can validate the results. Another attempt is in the Bengali language that uses document summary pairs from printed NCTB books[29] but does not release the corpus publicly. The sole corpus for the Dogri language is also not public[32]. A corpus consisting of 71 folktales is the sole Konkani corpus[30] and has not been released publicly. A work on Sanskrit text summarization uses Wikipedia articles for the task[28]. However, the dataset is also not available publicly. A work on Kannada text summarization uses IR-based approaches but does not give details of the dataset used[31]. Overall, most if not all works on Indian lan- guage summarization do not have a public dataset and the works can not be substantiated by any studies that are independent of the original research papers. 3. Task Definition The ILSUM task is a classic automatic summarization task where given a news article the par- ticipants are expected to generate a meaningful summary for the article. The summary can be either extractive or abstractive in nature. Traditionally the summarization tasks have been focused on generating a fixed-length summary irrespective of the input article length. This was especially the case with the DUC[34] and TAC[35] tasks and has since continued for a ma- jority of the summarization tasks elsewhere. However, unlike DUC and TAC datasets where the length of the source articles and human generated summaries were controlled, this is not the case with more recent large scale corpora. If the source articles vary in length and infor- mational content and so do the human summaries, forcing a fixed-length summary makes less sense. Keeping this in mind we propose a different approach and do not attempt to generate a fixed-length summary. Instead, participants are expected to predict an appropriate summary length for each article and we only limit the maximum summary length to 75 words. We argue that too long or short length summary compared to the ground truth summary will adversely affect ROUGE precision or recall respectively and the F-measure will implicitly be penalized. For this task we use standard ROUGE metrics Rouge-1, Rouge-2 and Rouge-4 F-scores are used for evaluation. To encourage participation and provide real time feedback a Kaggle like submission plat- form was provided to the participants. A separate leaderboard was provided for each language. During the validation phase, participants could submit runs on a blind validation dataset and instantly get the rouge scores. The leaderboard would display the highest score for each team along with the run id. During the test phase, participants could submit a maximum of three runs on the test data and see the rouge metrics instantly like in validation phase. The submis- sion platform is shown below in figure 1 Figure 1: ILSUM Submission Platform 4. Dataset The dataset for this task is built using articles and headline pairs from several leading news- papers in the country. We have provided 10,000+ news articles for Hindi, 12000+ articles for Gujarati and 17900+ articles for Indian English. Table 1 shows the detail statistics of the train, test, and validation dataset. The task is to generate a meaningful fixed-length summary, either extractive or abstractive, for each article. While several previous works in other languages use news articles - headlines pair, the current dataset poses a unique challenge of code-mixing and script mixing. It is very common for news articles to borrow phrases from English, even if the article itself is written in an Indian Language. Examples like those shown below are a common occurrence both in the headlines as well as in the articles. • Gujarati: ”IND vs SA, 5મી T20 તસવીરોમાં: વરસાદે િવલન બની મજા બગાડી” (India vs SA, 5th T20 in pictures: rain spoils the match) • Hindi: ”LIC के IPO में पैसा लगाने वालों का टू टा िदल, आई एक और नुकसानदेह खबर” (Investors of LIC IPO left broken hearted, yet another bad news) 4.1. Dataset Creation The news for ILSUM were scraped from the following news sites: • www.indiatvnews.com(English) • https://www.indiatv.in(Hindi) • https://www.divyabhaskar.co.in(Gujarati) • https://gujarati.news18.com(Gujarati) The data was collected using a combination of web scraping tools beautifulsoup and Oc- toparse. We initially collected 19,839 English, 22,349 Gujarati, and 11,750 Hindi URLs. Next, we cleaned the data by removing the HTML codes and any additional junk like extra spaces. Further, we dropped the articles where the headlines were too short. Only articles where head- line lenght was atleast 20 words were retained. The final corpus size is as shown in table 1 We assigned a unique id for each data record collected by computing a hash using the heading of the articles which are unique. The dataset was divided into train, test, and validation of size 70%(Train), 25%(Test) and 5%(Validation) respectively. Table 1 Dataset Distribution Hindi Gujarati English Training Set 6962 8460 12565 Validation Set 569 605 899 Test Set 2842 3021 4487 Total 10373 12086 17951 More details about the data are provided in table 2 below. The table contains number of sen- tences and words per article and per headline for all the three languages. It also shows number of codemixed articles (C.M.A.) and codemixed summaries(C.M.S.) for hindi and gujarati. As ev- ident, english documents are the longest (in number of words), followed Hindi while Gujarati documents are the shortest. On the other hand, headlines are the longest in Hindi articles fol- lowed by English and Gujarati. There is a much higher level of codemixing in Gujarati articles compared to Hindi articles. Table 2 Corpus Statistics Hindi Gujarati English Train Val Test Train Val Test Train Val Test Sents/Article 17.13 17.78 17.73 23.41 23.78 22.99 19.23 19.54 19.58 Words/Article 407.41 422.68 421.45 369.35 375.74 364.73 487.33 494.76 495.98 Sents/Summary 1.6 1.61 1.63 1.29 1.17 1.18 1.27 1.3 1.28 Words/Summary 37 36.85 37.44 29.06 29.41 28.75 33.39 33.62 33.42 C.M.A. 248 30 125 286 32 91 C.M.S. 363 25 100 2804 206 1011 5. Methodology In this section, we briefly discuss the approaches used by ILSUM 2022 participants. A majority of the teams preferred using large pre-trained models like BART, Pegasus, etc. for summa- rization and only a few approaches used traditional unsupervised methods. Notably, except for language-specific pre-trained models, none of the teams used any or language-specific re- sources, not even a stemmer or stopword list. This is counterintuitive in a task that would benefit widely from using linguistic resources. One possible reason is the lack of easy avail- ability of such resources. Unlike for English, a limited number of resources exist for Hindi or Gujarati most of which are not well evaluated. This also gives us a pointer for the next ver- sion of ILSUM, which is to make these resources easily accessible and encouraging teams to use them. The summary of systems used by different teams for Hindi, Gujarati and English is described in table 3, 4 and 5 • MT-NLP IIIT-H[36]: Team MT-NLP-IITH achieved best performance in all three sum- marization tasks. The authors used various transformer models by fine-tuning and con- sidering text summarization as a bottleneck task. For Hindi and Gujarati MT5, MBart, and IndicBART were finetuned for five epochs with a learning rate 5e-5 and max input length 512. Where best-performing model for Hindi is MT5 while MBart performed best for Gujarati. For English, PEGASUS, BART, T5 and ProphetNet were finetuned with similar hyperparameters, and PEGASUS outperformed other models on text data. • HakunaMatata[37]: mT5 and IndicBART are fine-tuned with actual and augmented data of size five times bigger than actual data. Fine-tuned IndicBART outperformed mT5 on all three tasks. • Next Gen NLP[38]: PEGASUS model worked best for this team on English and Gujarati where they use translation mapping-based approach. For hindi they used fine-tuned IndicBART model with augmented data. • PICT CL Lab[39]: This team used a transformer-based abstract summary generation approach by Indic-BART based model, fine-tuned using language modelling loss. • TextSumEval[40]: After preprocessing by removing multiple punctuations and emoti- cons, this team conducted four different experiments using LSTM, BART, GPT and T5 transformer, and T5 model achieved the best result for this team on English task. • SUMIL22[41]: is one of the teams that use approaches other than pretrained LLMs. They calculate various text features such as sentence position, sentence length, sentence similarity, frequent words, and sentence numbers for each sentence. These text features and their optimized weights are used for sentence ranking, and then the summary is generated by selecting top-ranked sentences. The weight optimization of text features is done using the population-based meta-heuristic approach, Genetic Algorithm (GA). • Summarize2022[42]: : For the English task, authors proposed a word frequency algorithm- based extractive text summarisation technique. Word frequency is calculated as the ratio of the frequency of a word and the frequency of the most occurring word in the text. Then sentence score is obtained by summing up the word frequency of all words occurring in a sentence. The mean of all sentence scores in the document is considered as a threshold to retain sentences in summary from the original text. • ILSUM_2022_SANGITA[43]: The author proposed encoder-decoder-based architec- ture for the summarization task. Encoder Bi-LSTM has a hidden state dimension = 128; decoder lstm has a hidden dimension = 256. The word embedding size = 300. model is trained using rmsprop optimiser with sparse categorical cross-entropy loss for 50 epochs with a learning rate of Bart and batch size of 32. • IIIT_Ranchi[44]: Extractive summarization approach using K means clustering was done by this team where clusters were created using sentence similarity scores. Where no of clusters for a document containing fifteen sentences is six, five for a document containing six sentences and a document containing less than six sentences were left unmodified. • SSNCSENLP[45]: mT5_m2m_CrossSum, a large-scale cross-lingual abstractive sum- marization model is used by this team to generate an abstractive summary. Table 3 Methodology used for Hindi Team Name Method Description MT-NLP IIIT-H[36] MT5, MBart, and IndicBART. best in MT5 HakunaMatata[37] MT5, and IndicBART with Data augmentation, best using IndicBART Next Gen NLP[38] Fine-tuned IndicBART Fine-tuned XL-Sum, best with IndicBART Fine-tuned mBART PICT CL Lab 2[39] Fine-tuned IndicBART IIIT_Ranchi[44] Extractive Summarization through K means clustring Table 4 Methodology used for Gujarati Team Name Method Description MT-NLP IIIT-H[36] MT5[6], MBart[7] and IndicBART. best in MBart HakunaMatata[37] MT5, and IndicBART with Data augmentation Next Gen NLP[38] Translation Mapping with PEGASUS, Fine-tuned mBART, Fine-tuned XL-Sum. best Translation Mapping with PEGASUS IIIT_Ranchi[44] Extractive Summarization through K means clustring Table 5 Methodology used on English Data Team Name Method Description MT-NLP IIIT-H[36] PEGASUS, BART, T5 and ProphetNet. PEGASUS gave best result Next Gen NLP[38] Fine-tuned PEGASUS Fine-tuned BRIO, SentenceBERT leveraged for summarization Fine-tuned T5 HakunaMatata[37] MT5, and IndicBART with Data augmentation TextSumEval[40] LSTM based sequence-to-sequence model, BART model, GPT model, and T5 model, best with T5 Model SUMIL22[41] a population-based meta heuristic approach Genetic Algorithm Summarize2022[42] Word Frequency Algorithm ILSUM_2022_SANGITA[43] Bi-LSTM based encoder and LSTM Based Decoder IIIT_Ranchi[44] Extractive Summarization through K means clustering 6. Results This section discusses results of runs submitted by different teams for all subtasks. Total of 12 teams submitted 47 runs across all subtasks. The summary of participation statistics is shown Table 6 Participation Statistics #Teams Registered #Teams Submitted Runs #Runs Submitted #Paper Submitted 56 12 47 10 in Table 6. Table 7, 8 and 9 shows the performance of best runs submitted by each team on Hindi, Gujarati and English tasks, respectively. Table 7 Performance of teams on Language summarization in Hindi F1 Score Rank Team Name ROUGE-1 ROUGE-2 ROUGE-3 ROUGE-4 1 MT-NLP IIIT-H[36] 0.607 0.510 0.484 0.471 2 HakunaMatata[37] 0.592 0.492 0.465 0.452 3 Euclido 0.583 0.480 0.452 0.439 4 Next Gen NLP[38] 0.556 0.455 0.427 0.414 5 PICT CL Lab 2[39] 0.544 0.443 0.419 0.400 6 IIIT_Ranchi[44] 0.327 0.174 0.136 0.126 Late Entry SSNCSENLP[45] 0.379 0.225 0.170 0.135 Table 8 Performance of teams on Language summarization in Gujarati F1 Score Rank Team Name ROUGE-1 ROUGE-2 ROUGE-3 ROUGE-4 1 MT-NLP IIIT-H[36] 0.261 0.165 0.138 0.124 2 HakunaMatata[37] 0.243 0.146 0.119 0.106 3 Euclido 0.225 0.123 0.091 0.075 4 Next Gen NLP[38] 0.209 0.119 0.095 0.084 5 IIIT_Ranchi[44] 0.176 0.085 0.063 0.053 Some of the summaries generated by the participating teams are listed alongside the gold- standard summaries below. Some of the summaries are codemixed and use one or two english words besides using english numerals. The quality of code-mixed summaries generated by the participating teams are at par with single script summaries. Hindi • Original: िहमाचल पर्देश: Flash Flood क वजह से नाले में अचानक बढ़ा पानी, 1 क मौत, 9 लापता",लाहौल स्पित के एसपी मानव वमार् ने बताया िक लाहौल स्पित क उदयपुर िडवीजन में फ्लैश फ्लड क वजह से 9 लोग लापता हैं। • MT-NLP IIIT-H : IANS द्ारा दी गई सूचना के अनुसार, आपदा मनाली-लेह राजमागर् पर स्थत उदयपुर में हई और तो जंग नदी पर एक महत्वपूणर् पुल क्षितगर्स्त हो गया। रपोट्सर् में कहा गया है िक पयर् टकों सिहत कई वाहन राजमागर् पर फंस गए हैं। Table 9 Performance of teams on Language summarization in English F1 Score Rank Team Name ROUGE-1 ROUGE-2 ROUGE-3 ROUGE-4 1 MT-NLP IIIT-H[36] 0.558 0.446 0.426 0.418 2 Euclido 0.558 0.444 0.422 0.414 3 Next Gen NLP[38] 0.557 0.443 0.421 0.412 4 HakunaMatata[37] 0.522 0.401 0.379 0.370 5 TextSumEval[40] 0.479 0.354 0.330 0.322 6 SUMIL22[41] 0.384 0.258 0.231 0.219 7 Team Crimson 0.362 0.228 0.201 0.190 8 Summarize2022[42] 0.340 0.208 0.181 0.171 9 ILSUM_2022_SANGITA[43] 0.328 0.166 0.122 0.098 10 IIIT_Ranchi[44] 0.321 0.186 0.163 0.154 • HakunaMatata: लाहौल स्पित के एसपी मानव वमार् ने बताया िक लाहौल स्पित क उदयपुर िडवीजन में फ्लैश फ्लड क वजह से 9 लोग लापता हैं। • Euclido: िहमाचल पर्देश के लाहौल स्पित में अचानक आई Flash Flood क वजह से तो जंग नाले (Tozing Nullah) में पानी बढ़ गया, जस वजह से एक व्यिक् क मौत हो गई, एक घायल है और 9 लोग लापता है। Gujarati • Original: Ahmedabad Airport: ધુળેટીના તહે વારને લઈ અમદાવાદ એરપોટર્ મેઘધનુષી રંગોમાં રંગાયું,"અમદાવાદના સરદાર વવૠભભાઈ પટે લ ઈન્ટરનેશનલ એરપોટર્ (Sardar Vallabhbhai Patel International Airport, Ahmedabad) પર રંગોત્સવ હોળી (Holi 20222) ના તહે વારને મનાવવાનો થનગનાટ ઉડીને આંખે વળગી રઽૠો છે . મુસાફરોને હોળી (Holi Festival) ના મેઘધનુષી રંગોમાં તરબોળ કરવા એરપોટર્ (Ahmedabad Airport) પ રસરમાં િવિવધ બળૠવૃિત્તઓનું આયોજન કરવામાં આવ્યું છે . • MT-NLP IIIT-H : મુસાફરોને વરલી અને માંડલ જેવી પરંપરાગત કળાના િવવધ સ્વરૂપોથી મા હત- ગાર કરવાનો બળૠયાસ કરવામાં આવ્યો છે . દેશ-િવદેશના બળૠવાસીઓ અહીં િવિવધ કળાને માણવાનો અનુભવ સેલ્ફીમાં પણ કે દ કરી રઽૠા છે . • HakunaMatata: SVPI એરપોટર્ પર ભારતીય કળાની વૈિવધ્યસભરતાનું આબેહૂબ િનદશર્ન કરા- વવામાં આવ્યું છે . મુસાફરો રંગબેરંગી માહોલમાં મંડલા અને વરલી જેવી પરંપરાગત કળાનો અદભૂત અનુભવ પણ માણી શકે તેવુ આયોજન કરવામાં આવ્યું છે . • Euclido: મુસાફરો રંગબેરંગી માહોલમાં મંડલા અને માંડલ જેવી પરંપરાગત કળાના િવવધ સ્વરૂ- પોથી મા હતગાર કરવાનો બળૠયાસ કરવામાં આવ્યો છે 7. Conclusion and Future Work The Indian Language Summarization (ILSUM) track at FIRE’22 is the first attempt to create benchmarked corpora for text summarization of Indian languages such as Hindi and Gujarati in addition to English. The majority of the summarization systems, submitted by the vari- ous participants, were based on pre-trained models like MT5, MBart, and IndicBART. Some of the participants also submitted systems using traditional unsupervised approaches, such as TexRank. The reported evaluation metric, the rouge F-Score, was comparable between English and Hindi corpora but significantly lower in Gujarati corpora. In the next edition of the ILSUM, we are planning to create a similar corpus for other languages like Bengali and Dravidian lan- guages like Tamil and Telugu. References [1] T. Wolf, L. Debut, V. Sanh, J. Chaumond, C. Delangue, A. Moi, P. Cistac, T. Rault, R. Louf, M. Funtowicz, J. Davison, S. Shleifer, P. von Platen, C. Ma, Y. Jernite, J. Plu, C. Xu, T. L. Scao, S. Gugger, M. Drame, Q. Lhoest, A. M. Rush, Transformers: State-of-the-art natural language processing, in: Q. Liu, D. Schlangen (Eds.), Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, EMNLP 2020 - Demos, Online, November 16-20, 2020, Association for Computational Linguistics, 2020, pp. 38–45. URL: https://doi.org/10.18653/v1/2020.emnlp-demos.6. doi:10.18653/ v1/2020.emnlp- demos.6 . [2] P. Mehta, T. Mandl, P. Majumder, S. Gangopadhyay, Report on the FIRE 2020 evaluation initiative, SIGIR Forum 55 (2021) 3:1–3:11. URL: https://doi.org/10.1145/3476415.3476418. doi:10.1145/3476415.3476418 . [3] T. Mandl, S. Modha, G. K. Shahi, H. Madhu, S. Satapara, P. Majumder, J. Schäfer, T. Ranas- inghe, M. Zampieri, D. Nandini, A. K. Jaiswal, Overview of the HASOC subtrack at FIRE 2021: Hatespeech and offensive content identification in english and indo-aryan languages, in: P. Mehta, T. Mandl, P. Majumder, M. Mitra (Eds.), Working Notes of FIRE 2021 - Forum for Information Retrieval Evaluation, Gandhinagar, India, December 13-17, 2021, volume 3159 of CEUR Workshop Proceedings, CEUR-WS.org, 2021, pp. 1–19. URL: http://ceur-ws.org/Vol-3159/T1-1.pdf. [4] T. Mandl, S. Modha, G. K. Shahi, A. K. Jaiswal, D. Nandini, D. Patel, P. Majumder, J. Schäfer, Overview of the HASOC track at FIRE 2020: Hate speech and offensive content identifi- cation in indo-european languages, in: P. Mehta, T. Mandl, P. Majumder, M. Mitra (Eds.), Working Notes of FIRE 2020 - Forum for Information Retrieval Evaluation, Hyderabad, India, December 16-20, 2020, volume 2826 of CEUR Workshop Proceedings, CEUR-WS.org, 2020, pp. 87–111. URL: http://ceur-ws.org/Vol-2826/T2-1.pdf. [5] S. Modha, T. Mandl, P. Majumder, D. Patel, Overview of the HASOC track at FIRE 2019: Hate speech and offensive content identification in indo-european languages, in: P. Mehta, P. Rosso, P. Majumder, M. Mitra (Eds.), Working Notes of FIRE 2019 - Fo- rum for Information Retrieval Evaluation, Kolkata, India, December 12-15, 2019, vol- ume 2517 of CEUR Workshop Proceedings, CEUR-WS.org, 2019, pp. 167–190. URL: http: //ceur-ws.org/Vol-2517/T3-1.pdf. [6] H. Madhu, S. Satapara, S. Modha, T. Mandl, P. Majumder, Detecting offensive speech in conversational code-mixed dialogue on social media: A contextual dataset and benchmark experiments, Expert Systems with Applications (2022) 119342. [7] S. Modha, P. Majumder, T. Mandl, C. Mandalia, Detecting and visualizing hate speech in social media: A cyber watchdog for surveillance, Expert Syst. Appl. 161 (2020) 113725. URL: https://doi.org/10.1016/j.eswa.2020.113725. doi:10.1016/j.eswa.2020.113725 . [8] M. Subramanian, R. Ponnusamy, S. Benhur, K. Shanmugavadivel, A. Ganesan, D. Ravi, G. K. Shanmugasundaram, R. Priyadharshini, B. R. Chakravarthi, Offensive language detection in tamil youtube comments by adapters and cross-domain knowledge trans- fer, Comput. Speech Lang. 76 (2022) 101404. URL: https://doi.org/10.1016/j.csl.2022.101404. doi:10.1016/j.csl.2022.101404 . [9] B. R. Chakravarthi, P. K. Kumaresan, R. Sakuntharaj, A. K. Madasamy, S. Thavareesan, B. Premjith, S. K, S. C. Navaneethakrishnan, J. P. McCrae, T. Mandl, Overview of the hasoc-dravidiancodemix shared task on offensive language detection in tamil and malay- alam, in: P. Mehta, T. Mandl, P. Majumder, M. Mitra (Eds.), Working Notes of FIRE 2021 - Forum for Information Retrieval Evaluation, Gandhinagar, India, December 13-17, 2021, volume 3159 of CEUR Workshop Proceedings, CEUR-WS.org, 2021, pp. 589–602. URL: http://ceur-ws.org/Vol-3159/T3-1.pdf. [10] S. Banerjee, K. Chakma, S. K. Naskar, A. Das, P. Rosso, S. Bandyopadhyay, M. Choudhury, Overview of the mixed script information retrieval (MSIR) at FIRE-2016, in: P. Majumder, M. Mitra, P. Mehta, J. Sankhavara, K. Ghosh (Eds.), Working notes of FIRE 2016 - Forum for Information Retrieval Evaluation, Kolkata, India, December 7-10, 2016, volume 1737 of CEUR Workshop Proceedings, CEUR-WS.org, 2016, pp. 94–99. URL: http://ceur-ws.org/ Vol-1737/T3-1.pdf. [11] P. Gupta, K. Bali, R. E. Banchs, M. Choudhury, P. Rosso, Query expansion for mixed- script information retrieval, in: S. Geva, A. Trotman, P. Bruza, C. L. A. Clarke, K. Järvelin (Eds.), The 37th International ACM SIGIR Conference on Research and Development in Information Retrieval, SIGIR ’14, Gold Coast , QLD, Australia - July 06 - 11, 2014, ACM, 2014, pp. 677–686. URL: https://doi.org/10.1145/2600428.2609622. doi:10.1145/ 2600428.2609622 . [12] P. Bhattacharya, K. Ghosh, S. Ghosh, A. Pal, P. Mehta, A. Bhattacharya, P. Majumder, Overview of the FIRE 2019 AILA track: Artificial intelligence for legal assistance, in: P. Mehta, P. Rosso, P. Majumder, M. Mitra (Eds.), Working Notes of FIRE 2019 - Forum for Information Retrieval Evaluation, Kolkata, India, December 12-15, 2019, volume 2517 of CEUR Workshop Proceedings, CEUR-WS.org, 2019, pp. 1–12. URL: http://ceur-ws.org/ Vol-2517/T1-1.pdf. [13] P. Bhattacharya, P. Mehta, K. Ghosh, S. Ghosh, A. Pal, A. Bhattacharya, P. Majumder, FIRE 2020 AILA track: Artificial intelligence for legal assistance, in: P. Majumder, M. Mitra, S. Gangopadhyay, P. Mehta (Eds.), FIRE 2020: Forum for Information Retrieval Evaluation, Hyderabad, India, December 16-20, 2020, ACM, 2020, pp. 1–3. URL: https://doi.org/10. 1145/3441501.3441510. doi:10.1145/3441501.3441510 . [14] V. Parikh, U. Bhattacharya, P. Mehta, A. Bandyopadhyay, P. Bhattacharya, K. Ghosh, S. Ghosh, A. Pal, A. Bhattacharya, P. Majumder, AILA 2021: Shared task on artificial in- telligence for legal assistance, in: D. Ganguly, S. Gangopadhyay, M. Mitra, P. Majumder (Eds.), FIRE 2021: Forum for Information Retrieval Evaluation, Virtual Event, India, De- cember 13 - 17, 2021, ACM, 2021, pp. 12–15. URL: https://doi.org/10.1145/3503162.3506571. doi:10.1145/3503162.3506571 . [15] V. Parikh, V. Mathur, P. Mehta, N. Mittal, P. Majumder, Lawsum: A weakly supervised approach for indian legal document summarization, CoRR abs/2110.01188 (2021). URL: https://arxiv.org/abs/2110.01188. arXiv:2110.01188. [16] S. Ghosh, A. Wyner, Identification of rhetorical roles of sentences in indian legal judg- ments, in: Legal Knowledge and Information Systems: JURIX 2019: The Thirty-second Annual Conference, volume 322, IOS Press, 2019, p. 3. [17] M. Amjad, G. Sidorov, A. Zhila, Data augmentation using machine translation for fake news detection in the urdu language, in: N. Calzolari, F. Béchet, P. Blache, K. Choukri, C. Cieri, T. Declerck, S. Goggi, H. Isahara, B. Maegaard, J. Mariani, H. Mazo, A. Moreno, J. Odijk, S. Piperidis (Eds.), Proceedings of The 12th Language Resources and Evalua- tion Conference, LREC 2020, Marseille, France, May 11-16, 2020, European Language Re- sources Association, 2020, pp. 2537–2542. URL: https://aclanthology.org/2020.lrec-1.309/. [18] M. Amjad, N. Ashraf, A. Zhila, G. Sidorov, A. Zubiaga, A. F. Gelbukh, Threat- ening language detection and target identification in urdu tweets, IEEE Access 9 (2021) 128302–128313. URL: https://doi.org/10.1109/ACCESS.2021.3112500. doi:10. 1109/ACCESS.2021.3112500 . [19] P. Mehta, P. Majumder, Optimum parameter selection for K.L.D. based authorship attribu- tion in gujarati, in: Sixth International Joint Conference on Natural Language Processing, IJCNLP 2013, Nagoya, Japan, October 14-18, 2013, Asian Federation of Natural Language Processing / ACL, 2013, pp. 1102–1106. URL: https://aclanthology.org/I13-1155/. [20] P. Mehta, P. Majumder, Large scale quantitative analysis of three indo-aryan languages, J. Quant. Linguistics 23 (2016) 109–132. URL: https://doi.org/10.1080/09296174.2015.1071151. doi:10.1080/09296174.2015.1071151 . [21] P. Mehta, From extractive to abstractive summarization: A journey, in: H. He, T. Lei, W. Roberts (Eds.), Proceedings of the ACL 2016 Student Research Workshop, Berlin, Ger- many, August 7-12, 2016, Association for Computational Linguistics, 2016, pp. 100–106. URL: https://doi.org/10.18653/v1/P16-3015. doi:10.18653/v1/P16- 3015 . [22] P. Mehta, P. Majumder, Effective aggregation of various summarization techniques, Inf. Process. Manag. 54 (2018) 145–158. URL: https://doi.org/10.1016/j.ipm.2017.11.002. doi:10. 1016/j.ipm.2017.11.002 . [23] S. Modha, P. Majumder, T. Mandl, R. Singla, Design and analysis of microblog-based summarization system, Social Network Analysis and Mining 11 (2021) 1–16. URL: https: //doi.org/10.1007/s13278-021-00830-3. [24] S. Sinha, G. N. Jha, An overview of indian language datasets used for text summa- rization, CoRR abs/2203.16127 (2022). URL: https://doi.org/10.48550/arXiv.2203.16127. doi:10.48550/arXiv.2203.16127 . arXiv:2203.16127. [25] R. Nallapati, B. Zhou, C. N. dos Santos, Ç. Gülçehre, B. Xiang, Abstractive text summa- rization using sequence-to-sequence rnns and beyond, in: Y. Goldberg, S. Riezler (Eds.), Proceedings of the 20th SIGNLL Conference on Computational Natural Language Learn- ing, CoNLL 2016, Berlin, Germany, August 11-12, 2016, ACL, 2016, pp. 280–290. URL: https://doi.org/10.18653/v1/k16-1028. doi:10.18653/v1/k16- 1028 . [26] B. Hu, Q. Chen, F. Zhu, LCSTS: A large scale chinese short text summarization dataset, in: L. Màrquez, C. Callison-Burch, J. Su, D. Pighin, Y. Marton (Eds.), Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, EMNLP 2015, Lisbon, Portugal, September 17-21, 2015, The Association for Computational Linguistics, 2015, pp. 1967–1972. URL: https://doi.org/10.18653/v1/d15-1229. doi:10.18653/v1/d15- 1229 . [27] E. S. Soriano, V. Ahuir, L. Hurtado, J. González, DACSA: A large-scale dataset for automatic summarization of catalan and spanish newspaper articles, in: M. Carpuat, M. de Marneffe, I. V. M. Ruíz (Eds.), Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL 2022, Seattle, WA, United States, July 10-15, 2022, Association for Computational Linguistics, 2022, pp. 5931–5943. URL: https://doi.org/10.18653/v1/2022. naacl-main.434. doi:10.18653/v1/2022.naacl- main.434 . [28] S. Barve, S. Desai, R. Sardinha, Query-based extractive text summarization for san- skrit, in: S. Das, T. Pal, S. Kar, S. C. Satapathy, J. K. Mandal (Eds.), Proceedings of the 4th International Conference on Frontiers in Intelligent Computing: Theory and Applications, FICTA 2015, Durgapur, India, 16-18 November 2015, volume 404 of Ad- vances in Intelligent Systems and Computing, Springer, 2015, pp. 559–568. URL: https: //doi.org/10.1007/978-81-322-2695-6_47. doi:10.1007/978- 81- 322- 2695- 6\_47 . [29] R. R. Chowdhury, M. T. Nayeem, T. T. Mim, M. S. R. Chowdhury, T. Jannat, Unsuper- vised abstractive summarization of bengali text documents, in: P. Merlo, J. Tiedemann, R. Tsarfaty (Eds.), Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume, EACL 2021, Online, April 19 - 23, 2021, Association for Computational Linguistics, 2021, pp. 2612–2619. URL: https: //doi.org/10.18653/v1/2021.eacl-main.224. doi:10.18653/v1/2021.eacl- main.224 . [30] J. D’Silva, U. Sharma, Development of a konkani language dataset for automatic text summarization and its challenges, International Journal of Engineering Research and Technology. International Research Publication House. ISSN (2019) 0974–3154. [31] V. R. Embar, S. R. Deshpande, A. Vaishnavi, V. Jain, J. S. Kallimani, saramsha-a kannada abstractive summarizer, in: 2013 International Conference on Advances in Computing, Communications and Informatics (ICACCI), IEEE, 2013, pp. 540–544. [32] S. Gandotra, B. Arora, Feature selection and extraction for dogri text summarization, in: Rising Threats in Expert Applications and Solutions, Springer, 2021, pp. 549–556. [33] R. Kabeer, S. M. Idicula, Text summarization for malayalam documents - an experience, in: International Conference on Data Science & Engineering, ICDSE 2014, Kochi, India, August 26-28, 2014, IEEE, 2014, pp. 145–150. URL: https://doi.org/10.1109/ICDSE.2014. 6974627. doi:10.1109/ICDSE.2014.6974627 . [34] A. Nenkova, Automatic text summarization of newswire: Lessons learned from the document understanding conference, in: M. M. Veloso, S. Kambhampati (Eds.), Pro- ceedings, The Twentieth National Conference on Artificial Intelligence and the Seven- teenth Innovative Applications of Artificial Intelligence Conference, July 9-13, 2005, Pitts- burgh, Pennsylvania, USA, AAAI Press / The MIT Press, 2005, pp. 1436–1441. URL: http://www.aaai.org/Library/AAAI/2005/aaai05-228.php. [35] H. T. Dang, K. Owczarzak, Overview of the TAC 2008 update summarization task, in: Pro- ceedings of the First Text Analysis Conference, TAC 2008, Gaithersburg, Maryland, USA, November 17-19, 2008, NIST, 2008. URL: https://tac.nist.gov/publications/2008/additional. papers/update_summ_overview08.proceedings.pdf. [36] A. Urlana, S. M. Bhatt, N. Surange, M. Shrivastava, Indian Language Summarization using Pretrained Sequence-to-Sequence Models, in: Working Notes of FIRE 2022 - Forum for Information Retrieval Evaluation, Kolkata, India, December 9-13, 2022, CEUR Workshop Proceedings, CEUR-WS.org, 2022. [37] D. Taunk, V. Varma, Summarizing Indian Languages using Multilingual Transformers based Models, in: Working Notes of FIRE 2022 - Forum for Information Retrieval Evalu- ation, Kolkata, India, December 9-13, 2022, CEUR Workshop Proceedings, CEUR-WS.org, 2022. [38] R. Tangsali, A. Pingle, A. Vyawahare, I. Joshi, R. Joshi, Implementing Deep Learning- Based Approaches for Article Summarization in Indian Languages, in: Working Notes of FIRE 2022 - Forum for Information Retrieval Evaluation, Kolkata, India, December 9-13, 2022, CEUR Workshop Proceedings, CEUR-WS.org, 2022. [39] A. Agarwal, S. Naik, S. Sonawane, Abstractive Text Summarization for Hindi Language using IndicBART, in: Working Notes of FIRE 2022 - Forum for Information Retrieval Evaluation, Kolkata, India, December 9-13, 2022, CEUR Workshop Proceedings, CEUR- WS.org, 2022. [40] S. Chakraborty, D. Kaushik, S. R. Laskar, P. Pakray, Exploring Text Summarization Mod- els for Indian Languages Summarization, in: Working Notes of FIRE 2022 - Forum for Information Retrieval Evaluation, Kolkata, India, December 9-13, 2022, CEUR Workshop Proceedings, CEUR-WS.org, 2022. [41] V. P. K. Doppalapudi, S. R. Srigadha, P. Verma, S. Pal, in: Working Notes of FIRE 2022 - Forum for Information Retrieval Evaluation, Kolkata, India, December 9-13, 2022, CEUR Workshop Proceedings, CEUR-WS.org, 2022. [42] N. Abinaya, S. Anbukkarasi, S. Varadhaganapathy, An Extractive Text Summarization Using Word Frequency Algorithm for English Text, in: Working Notes of FIRE 2022 - Forum for Information Retrieval Evaluation, Kolkata, India, December 9-13, 2022, CEUR Workshop Proceedings, CEUR-WS.org, 2022. [43] S. Singh, J. P. Singh, A. Deepak, Deep Learning based Abstractive Summarization for English Language, in: Working Notes of FIRE 2022 - Forum for Information Retrieval Evaluation, Kolkata, India, December 9-13, 2022, CEUR Workshop Proceedings, CEUR- WS.org, 2022. [44] K. Kumari, R. Kumari, An Extractive Approach for Automated Summarization of Indian Languages using Clustering Techniques, in: Forum for Information Retrieval Evaluation (Working Notes) (FIRE), CEUR-WS.org, 2022. [45] A. Krishnakumar, F. Naushin, M. KL, B. B, Text summarization for Indian languages using pre-trained models, in: Working Notes of FIRE 2022 - Forum for Information Retrieval Evaluation, Kolkata, India, December 9-13, 2022, CEUR Workshop Proceedings, CEUR- WS.org, 2022.