<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>MSD: Multilingual Sarcasm Detection using Deep Learning-Based Model</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Ranjeet Kumar</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Abhinav Kumar</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science and Engineering, Motilal Nehru National Institute of Technology Allahabad</institution>
          ,
          <addr-line>Prayagraj, Uttar Pradesh, 211004</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>Sarcasm detection is a crucial task in natural language processing (NLP), where the intended meaning of a statement diverges from its surface-level interpretation. Sarcasm detection plays a vital role in sentiment analysis and opinion mining. Historically, research in this domain has been limited to single-language input text. However, with the rise of social media, there has been a surge in multilingual data, where users express themselves in a mix of languages such as Hinglish, Tamil, and Malayalam. This paper addresses the need for multilingual sarcasm detection by presenting deep learning-based architectures such as BERT and Xlm-RoBERTa. The multilingual Tamil-English and Malayalam-English texts are first translated into their corresponding English text, and then BERT and Xlm-RoBERTa are used to fine-tune their weight for sarcasm identification. The proposed approach achieves promising performance with macro 1-score of 0.72 for both BERT and Xlm-RoBERTa models in the case of Tamil-English posts. In contrast, it achieves macro 1-scores of 0.71 for BERT and 0.73 for Xlm-RoBERTa in the case of Malayalam-English posts.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Sarcasm</kwd>
        <kwd>Multilingual</kwd>
        <kwd>Xlm-RoBERTa</kwd>
        <kwd>BERT</kwd>
        <kwd>NLP</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Sarcasm is familiar on social media sites, where a message’s deeper significance frequently difers from
its apparent reading. Sarcasm detection is a significant dificulty in many applications, mainly where
it’s essential to discern the speaker’s true viewpoint, like in discussion forums, customer evaluations,
and sentiment analysis tools [
        <xref ref-type="bibr" rid="ref1 ref2 ref3">1, 2, 3</xref>
        ]. Traditionally, lexical clues and certain language components have
been used to approach sarcasm detection as a text categorization problem. However, sarcasm detection
has become a far more dificult process with the emergence of social media data from platforms such as
Instagram and Twitter [
        <xref ref-type="bibr" rid="ref4 ref5 ref6">4, 5, 6</xref>
        ]. Comprehending sarcasm takes more than just recognizing linguistic
patterns; it frequently calls for prior knowledge, contextual awareness, or even a particular degree
of maturity or intelligence. Sarcasmic content abounds on social media, especially on microblogging
platforms, so it is impracticable to detect sarcastic content manually. Because of this, complex algorithms
that can automatically identify and understand sarcasm in user-generated content are now required.
Since sarcasm is based on intense emotions that come from one’s situation, attitude, or relationships
with other people, it frequently crosses over into emotion. People’s physical and psychological reactions
to emotions like love, hate, or fear are mirrored in online communication. Thus, emotion identification
in reviews or comments on social media has become an important part of several study fields [
        <xref ref-type="bibr" rid="ref10 ref7 ref8 ref9">7, 8, 9, 10</xref>
        ],
including sarcasm detection. Sarcastically worded content’s emotions can have a big impact on how
users understand user opinions, whether for businesses evaluating customer feedback or for online
platforms keeping an eye on public discourse [
        <xref ref-type="bibr" rid="ref11 ref12">11, 12</xref>
        ].
      </p>
      <p>
        Dealing with content that lacks obvious context, like headlines, is one of the trickiest parts of sarcasm
identification. Single-line news headlines are especially challenging to understand, in contrast to tweets
or longer social media posts that ofer surrounding information for sarcasm identification. It is possible
to significantly change the intended meaning of a sarcastic headline by mistaking it for a real statement.
This might result in poorly informed decisions in commercial applications or incorrect interpretations of
important information. This problem also highlights worries about how sarcasm could spread dangerous
content, including disparaging particular races or ethnic groups in ostensibly innocent yet satirical
headlines. Tools that can reliably identify sarcasm in this kind of information are essential to reducing
these dangers [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ].
      </p>
      <p>Automatic sarcasm identification has become more important and complex due to the increased
occurrence of multimodal and multilingual data, particularly as users combine languages like Hinglish,
Tamil, and Malayalam more frequently. Furthermore, one of the biggest challenges facing conventional
sarcasm detection techniques is the absence of context in headlines and short social media posts. To
tackle these challenges, this work explores the usability of two diferent deep learning models: (i) BERT
and (ii) Xlm-RoBERTa. To fine-tune these models, Tamil-English and Malayalam-English datasets are
ifrst translated into their corresponding English text.</p>
      <p>The rest of the paper is organized as follows: Section 2 lists related literature for identifying sarcastic
social media posts. Section 3 discusses the proposed methodology, results of the proposed model are
listed in Section 4, and Section 5 concludes the paper.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Literature Review</title>
      <p>
        The identification of sarcasm is a challenging topic in natural language processing (NLP) that has
attracted a lot of study interest since it relies on subtle linguistic clues such as context, exaggeration,
and incongruity. Recognizing sarcasm is essential in several applications, including sentiment analysis,
opinion mining, and social media content regulation. A variety of techniques have been investigated
to increase the accuracy of sarcasm recognition models [
        <xref ref-type="bibr" rid="ref14 ref15 ref16">14, 15, 16</xref>
        ]. This is especially correct when
working with multimodal data; here, the user can express himself by a combination of speech, text, and
images.
      </p>
      <sec id="sec-2-1">
        <title>2.1. Conventional Methods</title>
        <p>Sarcasm Detection through Sentiment and Incongruity: Detection of sarcasm using sentiment
analysis is a well-known research on sarcasm detection that uses a model that includes sentiment and
data to identify sarcasm. This architecture covered representations that are sentiment and
incongruityoriented to this configuration. By using the most used dataset on sarcasm detection, This study shows
some of the main shortcomings of other used models in the literature, especially in compiling big texts,
and emphasizes the need to include sentiment and sarcasm detection. [17]</p>
        <p>Multimodal Sarcasm Detection: Detection of sarcasm that can be handled with multimodal data,
photos, text, and audio is more important as the quantity of data generated from social media platforms
such as Instagram and Twitter keeps growing. The Multimodal Learning framework is one of the most
used methods in this area. By aligning text with visual and oral modalities, this framework enhances
the extraction of contextual and emotional elements using a cross-modal target attention mechanism.
In the case of the previous modal that prioritizes shared properties, the above framework indicates and
shows both the shared and diferent elements of sentiment analysis and detection of sarcasm. This
modal uses the junction between sentiment analysis and sarcasm detection to improve the model’s
capability to grasp task-specific distinctions. The efectiveness of this multimodal data in improving the
detection of sarcasm was explained by extensive testing of the MIL model on the MUStARD datasets,
which is a significant gain over what is already in place [18].</p>
        <p>Deep Learning Approaches with Contextual Features: In the case of social media datasets on
sarcasm identification, especially tweets, the use of contextual feature extraction in diferent conjunctions
using diferent deep-learning methods was reported. The writer wrote a system that collects manually
created contextual properties based on linguistics with a Convolutional Neural Network(CNN) to extract
features. Word embeddings were completed using FastText elements, which combined verbs, pronouns,
and event words to enhance contextual awareness. Using diferent machine learning classifiers to
compare the accuracy of the model, the author found that this model most outperformed other models
on the same datasets in terms of F1-score. To detect sarcasm, deep learning combined with contextual
awareness is the most used method, providing feature complexity and model accuracy [19].</p>
        <p>
          Sarcasm Detection in Code-Mixed Conversations: Bedi et al. [20] completed a noticeable
investigation focusing on the complexities associated with sarcasm identification and classification in
code-mixed languages, especially those that contain Hindi and English. The researcher demonstrated
an MSH-COMICS model and presented MaSaC, a unique multimodal dataset for the classification of
sarcasm identification. This architecture can analyze the complexities of code-mixed datasets.[
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] This
uses a tree structural approach that concentrates on the basis of sentences and diferent dialog. On the
basis of this experiment, MSH-COMICS accuracy is improved than previous models by obtaining a
higher F1-score value of classification and categorization of sarcasm identification. [20]
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Proposed Methodology</title>
      <p>The overall flow diagram of the proposed methodology can be seen in Figure 1.</p>
      <sec id="sec-3-1">
        <title>3.1. Dataset Description</title>
        <p>
          Malayalam-English and Tamil-English text data are used for this experiment. This dataset has social
media posts and their corresponding labels, sarcastic and non-sarcastic [21]. The characteristics of
datasets include casual conversations, specifically in Tamil and Malayalam, film industries, and cinemas
in south India. For example, “Rajapapanod pokan para . . . njangalkku njangade stephen chettan undallo
” and “Ee pattinum dislike adicha aalkkarod anik puncham mathram ” these types of sentences are
present in this dataset that shows richness in covering sarcasm in diferent forms, making it a more
worthy resource for developing and filtering sarcasm identification algorithms in low resource languages.
languages [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. The overall statistics of the datasets can be seen in Table 1. This dataset is separated
into training, testing and validation sets for the Malayalam and Tamil languages. The Validation data
for Malayalam has 2826 samples with 2305 non-sarcastic labels and 521 sarcastic labels. There are 13188
samples in the training data, with 2499 sarcastic and 10689 non-sarcastic labels. In testing datasets there
are no labels presents in this testing dataset there are 2826 rows in it [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]. In the validation dataset of
Tamil there are 636 samples which have 4630 non-sarcastic labels and 1706 sarcastic labels. In training
datasets there 29570 samples which have 7830 sarcastic labels and 21740 non-sarcastic labels. In case of
testing datasets there are 6338 rows are presents without any labels. This section indicates that the
model are fairly trained and evaluated for both the languages.
        </p>
        <p>The GoogleTranslator1 module from the deep translator package was used to translate the dataset.
The translator was configured to convert the text into English automatically after identifying the source
language, which may be either Tamil or Malayalam code-mixed text.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Translation Process</title>
        <p>The dataset used in this research consists of Tamil-English and Malayalam-English code-mixed text,
which needed to be translated into English for further analysis. To achieve this, we employed the
GoogleTranslator module from the deep translator package. This tool was selected for its ability to
automatically detect the source language, whether Tamil or Malayalam, and convert the text into
English. Given the nature of the dataset, which contains a wide range of casual conversations and social
media posts, accurately capturing linguistic nuances, especially sarcasm, was crucial. The automatic
translation process ofered a fast and eficient way to process the large dataset, ensuring consistency
in handling the initial translations. However, automatic translation tools are not without limitations,
particularly when dealing with sarcasm, informal language, and code-mixed text. To address this, we
implemented a secondary step in the translation process, where human validation was carried out
by a team of five language experts. Each team member is proficient in both the source languages
(Tamil and Malayalam) and English, allowing them to critically assess the quality and accuracy of
the machine-generated translations. During this validation phase, the team thoroughly reviewed the
translated text, comparing it to the original to identify any mistranslations, contextual inaccuracies,
or issues where the automatic translator failed to capture the correct meaning. If any discrepancies
were found, the team members manually corrected the translations to ensure they reflected the original
intent, especially for detecting sarcasm, which often involves subtle or indirect cues. By combining
machine translation with expert human validation, we ensured that the translations were both accurate
and contextually appropriate, making the dataset highly reliable for developing and testing sarcasm
identification algorithms in low-resource languages.</p>
        <sec id="sec-3-2-1">
          <title>Malayalam-English social media posts</title>
        </sec>
        <sec id="sec-3-2-2">
          <title>Dravidian Social</title>
        </sec>
        <sec id="sec-3-2-3">
          <title>Media Posts</title>
          <p>Tamil-English social
media posts
r
o
t
a
l
s
n
a
r
T</p>
        </sec>
        <sec id="sec-3-2-4">
          <title>English Translated</title>
        </sec>
        <sec id="sec-3-2-5">
          <title>Malayalam-English social media posts</title>
        </sec>
        <sec id="sec-3-2-6">
          <title>English Translated</title>
        </sec>
        <sec id="sec-3-2-7">
          <title>Tamil-English social media posts</title>
        </sec>
        <sec id="sec-3-2-8">
          <title>BERT</title>
        </sec>
        <sec id="sec-3-2-9">
          <title>Xlm-RoBERTa</title>
        </sec>
        <sec id="sec-3-2-10">
          <title>Not-sarcastic Sarcastic</title>
        </sec>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Model Selection and Training</title>
        <p>Two diferent deep learning-based models such as BERT and Xlm-RoBERTa were fine-tuned on the
English translated Tamil-English and Malayalam-English to classify it into sarcastic and not-sarcastic
classes.</p>
        <p>• BERT: BERT (Bidirectional Encoder Representations from Transformers) [22] is a deep learning
model designed by Google that has achieved state-of-the-art performance on a wide variety
of Natural Language Processing (NLP) tasks, including text classification. BERT has several
advantages: (i) Pre-trained contextualized embeddings: BERT can handle long-range dependencies
and understands words in the context of the entire sentence; (ii) Transfer learning: Pre-training
on a massive dataset means that fine-tuning on specific tasks requires significantly less data,
(iii) State-of-the-art performance: BERT has achieved top results in many text classification
benchmarks. Therefore, this work uses this model to fine-tuned with the translated Tamil-English
and Malayalam-English social media posts.
• XLM-RoBERTa: XLM-RoBERTa is an extension of the original BERT (Bidirectional Encoder
Representations from Transformers) and RoBERTa (Robustly Optimized BERT Pretraining Approach)
models. It leverages the strengths of RoBERTa while focusing on multilingual
understanding by being pre-trained on text in 100 languages. Unlike previous multilingual models like
mBERT (Multilingual BERT), XLM-RoBERTa doesn’t rely on language-specific tokens, making
it a language-agnostic model. There are several advantages of using Xlm-RoBERTa model for
text classification: (i) Multilingual support: XLM-R is specifically designed for multilingual tasks,
making it highly suitable for text classification in multiple languages. (ii) Cross-lingual transfer:
The model can be fine-tuned on one language and perform well on other languages, even with
little or no training data for the target language. (iii) Handling of low-resource languages: XLM-R
performs well even in languages with limited training data because of its extensive multilingual
pre-training. (iv) Contextual understanding: Like BERT and RoBERTa, XLM-R understands
words in the context of the entire sentence, providing rich, contextualized word embeddings
for classification tasks. Due to the robustness of Xlm-RoBERTa model, this paper utilized it and
ifne-tuned it on the translated English social media posts.</p>
        <p>We begin by preprocessing the data, where we initialize the Ktrain text transformer tailored to the
chosen model. Each input text is capped at 30 tokens to ensure consistency. Both the text inputs
and corresponding labels are transformed into the model’s expected format for training. The selected
pre-trained Transformer model is fine-tuned using the Adam optimizer, configured for classification
tasks. The model is trained for 50 epochs with a learning rate of 5 × 10− 5 and a batch size of 32.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Results</title>
      <p>The performance of XLM-RoBERTa-Base and BERT-Base-Multilingual models fine-tuned on
TamilEnglish and Malayalam-English can be seen in Table 2. In terms of identifying non-sarcastic posts,
both models performed admirably, with precision and 1-scores continuously surpassing 0.80 in both
languages. Both XLM-RoBERTa-Base and BERT-Base-Multilingual achieved 1-scores of 0.86 for
nonsarcastic postings in the Tamil-English dataset. However, both models had trouble distinguishing
sarcasm; their 1-scores dropped to about 0.57 and 0.58, respectively, indicating how dificult it is to
spot sarcastic content in code-mixed data. XLM-RoBERTa-Base fared marginally better than
BERTBase-Multilingual for the Malayalam-English dataset. This was especially true in the sarcastic class,
where XLM-Roberta-Base obtained an 1-score of 0.55 whereas BERT achieved an 1-score of 0.52.
This slight discrepancy implies that XLM-Roberta-Base is more appropriate for sarcasm detection in
postings written in Malayalam and English. With a weighted average 1-score of 0.84 against 0.83
for BERT-Base-Multilingual, XLM-RoBERTa-Base outperformed the other model in terms of overall
performance as measured by their weighted 1-scores. The confusion matrix and ROC curve of the
1.0
0.2
1.0
0.8
e
t
a
iiteR0.6
v
s
o
P
reu0.4
T
0.2
Confusion matrix</p>
      <p>Receiver operating characteristic curve
Confusion matrix</p>
      <p>Receiver operating characteristic curve
Non-sarcastic
Xlm-RoBERTa model for the Tamil-English language can be seen in Figures 2 and 3, respectively.
Similarly, the confusion matrix and ROC curve for the Xlm-RoBERTa model for the Malayalam-English
language can be seen in Figures 4 and 5, respectively.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>In this work, we assessed two cutting-edge multilingual transformer models for sarcasm detection
using code-mixed datasets in Tamil-English and Malayalam-English: BERT-Base-Multilingual and
XLM-Roberta-Base. With F1-scores continuously above 0.80, both models showed good performance in
identifying non-sarcastic text in both languages. The models’ dificulties with the sarcastic class, on
the other hand, were evident in their lower 1-scores (0.52 to 0.58), which underscores how dificult
it is to detect sarcasm in code-mixed data. For the Tamil-English dataset, both models fared similarly.
However, XLM-Roberta-Base did slightly better, especially for the Malayalam-English dataset, where
it beat Bert-Base-Multilingual in both the non-sarcastic and sarcastic classes. In comparison to
BertBase-Multilingual, which obtained a weighted average F1-score of 0.83 for Malayalam-English, the
XLM-Roberta-Base model demonstrated a stronger overall capacity to handle code-mixed data. These
ifndings imply that although transformer models are helpful in identifying non-sarcastic content,
more efort is necessary to enhance the detection of sarcasm, especially in low-resource, code-mixed
languages. Future work may investigate better identifying sarcasm in such challenging datasets by
integrating more contextual features such as multimodal data or sophisticated fine-tuning methods.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>The author(s) have not employed any Generative AI tools.
R. Ponnusamy, C. Subalalitha, B. R. Chakravarthi, Findings of shared task on sarcasm identification
in code-mixed dravidian languages, FIRE 2023 16 (2023) 22.
[17] A. Bhat, A. Chauhan, A deep learning based approach for multimodal sarcasm detection, in: 2022
4th International Conference on Advances in Computing, Communication Control and Networking
(ICAC3N), IEEE, 2022, pp. 2523–2528.
[18] A. Kumar, V. T. Narapareddy, V. A. Srikanth, A. Malapati, L. B. M. Neti, Sarcasm detection using
multi-head attention based bidirectional lstm, Ieee Access 8 (2020) 6388–6397.
[19] M. S. Razali, A. A. Halin, N. M. Norowi, S. C. Doraisamy, The importance of multimodality in
sarcasm detection for sentiment analysis, in: 2017 IEEE 15th Student Conference on Research and
Development (SCOReD), IEEE, 2017, pp. 56–60.
[20] M. Bedi, S. Kumar, M. S. Akhtar, T. Chakraborty, Multi-modal sarcasm detection and humor
classification in code-mixed conversations, IEEE Transactions on Afective Computing 14 (2021)
1363–1375.
[21] B. R. Chakravarthi, S. N, B. B, N. K, T. Durairaj, R. Ponnusamy, P. K. Kumaresan, K. K.</p>
      <p>Ponnusamy, C. Rajkumar, Overview of sarcasm identification of dravidian languages in
DravidianCodeMix@FIRE-2024, in: Forum of Information Retrieval and Evaluation FIRE - 2024,
DAIICT , Gandhinagar, 2024.
[22] A. Vaswani, Attention is all you need, Advances in Neural Information Processing Systems (2017).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>R.</given-names>
            <surname>Pandey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Tripathi</surname>
          </string-name>
          ,
          <article-title>Hybrid attention-based long short-term memory network for sarcasm identification</article-title>
          ,
          <source>Applied Soft Computing</source>
          <volume>106</volume>
          (
          <year>2021</year>
          )
          <fpage>107348</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>R.</given-names>
            <surname>Pandey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Tripathi</surname>
          </string-name>
          ,
          <article-title>A hybrid convolutional neural network for sarcasm detection from multilingual social media posts</article-title>
          ,
          <source>Multimedia Tools and Applications</source>
          (
          <year>2024</year>
          )
          <fpage>1</fpage>
          -
          <lpage>29</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <article-title>Hope speech detection in youtube comments</article-title>
          ,
          <source>Social Network Analysis and Mining</source>
          <volume>12</volume>
          (
          <year>2022</year>
          )
          <fpage>75</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Shi</surname>
          </string-name>
          ,
          <article-title>A prior probability of speaker information and emojis embedding approach to sarcasm detection</article-title>
          , in: 2021 IEEE International Conference on Engineering, Technology &amp;
          <string-name>
            <surname>Education (TALE),</surname>
            <given-names>IEEE</given-names>
          </string-name>
          ,
          <year>2021</year>
          , pp.
          <fpage>1033</fpage>
          -
          <lpage>1038</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Sripriya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bharathi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Nandhini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. Chinnaudayar</given-names>
            <surname>Navaneethakrishnan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Durairaj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Kumaresan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Rajkumar, Overview of the shared task on sarcasm identification of Dravidian languages (Malayalam and Tamil) in DravidianCodeMix, in: Forum of Information Retrieval and Evaluation FIRE -</article-title>
          <year>2023</year>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Sripriya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bharathi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Nandhini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. C.</given-names>
            <surname>Navaneethakrishnan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Durairaj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Kumaresan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Rajkumar, Overview of the shared task on sarcasm identification of dravidian languages (malayalam and tamil) in dravidiancodemix, in: Forum of Information Retrieval and Evaluation FIRE-</article-title>
          <year>2023</year>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>S.</given-names>
            <surname>Saumya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <article-title>Filtering ofensive language from multilingual social media contents: A deep learning approach</article-title>
          ,
          <source>Engineering Applications of Artificial Intelligence</source>
          <volume>133</volume>
          (
          <year>2024</year>
          )
          <fpage>108159</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Kumari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <article-title>Ja-nlp@ lt-edi-2023: Empowering mental health assessment: A robertabased approach for depression detection</article-title>
          ,
          <source>in: Proceedings of the Third Workshop on Language Technology for Equality, Diversity and Inclusion</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>89</fpage>
          -
          <lpage>96</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kumari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Pradhan</surname>
          </string-name>
          ,
          <article-title>Explainable deep learning for mental health detection from english and arabic social media posts</article-title>
          ,
          <source>ACM Transactions on Asian and Low-Resource Language Information Processing</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Saumya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <article-title>Detecting dravidian ofensive posts in miot: A hybrid deep learning framework</article-title>
          ,
          <source>ACM Trans. Asian Low-Resour. Lang. Inf. Process</source>
          . (
          <year>2023</year>
          ). URL: https: //doi.org/10.1145/3592602. doi:
          <volume>10</volume>
          .1145/3592602, just Accepted.
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>T. S.</given-names>
            <surname>Apon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Anan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. A.</given-names>
            <surname>Modhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Suter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I. J.</given-names>
            <surname>Sneha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. G. R.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <article-title>Banglasarc: A dataset for sarcasm detection</article-title>
          ,
          <source>in: 2022 IEEE Asia-Pacific Conference on Computer Science and Data Engineering (CSDE)</source>
          , IEEE,
          <year>2022</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>5</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Roy</surname>
          </string-name>
          ,
          <article-title>An advanced learning approach for detecting sarcasm in social media posts: Theory and solutions</article-title>
          ,
          <source>Social Science Quarterly</source>
          <volume>105</volume>
          (
          <year>2024</year>
          )
          <fpage>1857</fpage>
          -
          <lpage>1874</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>J. A.</given-names>
            <surname>Josephine</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. K.</given-names>
            <surname>Maharana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A. A.</given-names>
            <surname>Walid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Thulasimani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. S.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Tiwari</surname>
          </string-name>
          ,
          <article-title>Hybrid particle swarm optimization with deep learning driven sarcasm detection on social media</article-title>
          ,
          <source>in: 2022 International Conference on Automation, Computing and Renewable Systems (ICACRS)</source>
          , IEEE,
          <year>2022</year>
          , pp.
          <fpage>687</fpage>
          -
          <lpage>693</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>R.</given-names>
            <surname>Ratnavel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. G.</given-names>
            <surname>Joshua</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Varsini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. A.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <article-title>Sarcasm detection in tamil code-mixed data using transformers</article-title>
          ,
          <source>in: International Conference on Speech and Language Technologies for Low-resource Languages</source>
          , Springer,
          <year>2023</year>
          , pp.
          <fpage>430</fpage>
          -
          <lpage>442</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>R.</given-names>
            <surname>Sivakumar</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. M. C</surname>
            ,
            <given-names>B. M. Jenefer</given-names>
          </string-name>
          ,
          <article-title>Identifying the type of sarcasm in dravidian languages using deep-learning models</article-title>
          .,
          <source>in: FIRE (Working Notes)</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>278</fpage>
          -
          <lpage>286</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>N.</given-names>
            <surname>Sripriya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Durairaj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Nandhini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bharathi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Rajkumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Kumaresan</surname>
          </string-name>
          ,
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>