<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Unmasking Sarcasm: Exploring mBERT+CNN and LSTM Models for Sarcasm Identification in Code-Mixed Tamil and Malayalam Texts</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Sonith D</string-name>
          <email>sonithksd@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kavya G</string-name>
          <email>kavyamujk@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>H L Shashirekha</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science, Mangalore University</institution>
          ,
          <addr-line>Mangalore, Karnataka</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Forum for Information Retrieval Evaluation</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>On social media platforms, users express their thoughts and emotions in diverse ways including sentiments, sarcasm, signs of depression, and expressions of hatredness. Sarcasm is a form of expression where the intended meaning is opposite to the literal meaning of the words used, often to convey mockery or irony. This can significantly alter the perceived sentiment of a message, making it challenging for Natural Language Processing (NLP) tasks such as opinion and sentiment analysis. Detecting sarcasm is crucial because it helps in accurately interpreting user sentiments and can enhance the efectiveness of automated systems in processing and responding to user-generated content. In this direction, “Sarcasm Identification of Dravidian Languages in DravidianCodeMix@FIRE-2024"- a shared task organized at Forum for Information Retrieval Evaluation (FIRE) 2024, invites the research community to address the challenges of sarcasm detection in code-mixed Dravidian languages (Tamil-English and Malayalam-English). To explore the strategies for sarcasm identification in Dravidian languages, in this paper, we - team MUCS, describe the models proposed for the shared task. Two distinct models: i) Long Short-Term Memory (LSTM) model - trained using Keras embeddings, and ii) mBERT+CNN model - a combination of Transfer Learning (TL) (fine-tuning Multilingual Bidirectional Encoder Representations from Transformers (mBERT)) and Deep Learning (DL) approach (Convolutional Neural Network (CNN)) for building classifier, are proposed for sarcasm identification. Further, to overcome the data imbalance issue in the dataset, text augmentation technique is explored using Contextual word embedding augmenter from Natural Language Processing Augmentation (NLPAug) library, to increase the number of samples in the minority class. Among the proposed models, mBERT+CNN model outperformed LSTM model with macro F1 scores of 0.74 and 0.72 for Tamil-English and Malayalam-English datasets securing 1st and 4th ranks respectively.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Sarcasm Detection</kwd>
        <kwd>Data Augmentation</kwd>
        <kwd>Transfer Learning</kwd>
        <kwd>Deep Learning</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Social networking sites have become a great source of user-generated textual content that lends itself
to engagement via likes, shares, comments, and discussion [
        <xref ref-type="bibr" rid="ref1 ref2 ref3">1, 2, 3</xref>
        ]. This textual content corresponds
to various topics such as hate speech, hope speech, fake news etc., and sarcasm is one among them.
Sarcastic comments often draw a contrast between the literal words expressed and their intended
implication. For instance, if someone says, "Oh, what a masterpiece!" after watching a poorly received
movie, the literal praise is meant sarcastically. In contrast, non-sarcastic comments are straightforward
and convey the speaker’s true sentiment. For example, saying, "The movie was really engaging", directly
communicates genuine praise without any hidden meaning. Sarcastic text poses a unique challenge
for text analysis due to its reliance on contextual and frequently nuanced indications [
        <xref ref-type="bibr" rid="ref4">4, 5</xref>
        ]. Satirical
remarks on social media have the power to skew the interpretation of a discussion or a piece of content,
which can have an impact on interactions and public opinion. Accurately detecting irony can boost the
efectiveness of sentiment analysis systems and improve the analysis of comments/posts collected from
social media. Separating sincere statements from satirical ones is also crucial for proper interpretation
and response in the context of social media monitoring and content management. Hence, detecting
sarcasm on social media platforms becomes crucial. However, sarcasm detection presents significant
challenges as sarcastic remarks often depend on contextual understanding, tone, and cultural nuances
that are dificult to capture with standard text processing techniques. The code-mixed nature of textual
content on social media platforms adds its share of challenges to detect sarcastic content.
      </p>
      <p>Many of the works on sarcasm detection focus on high-resource languages, leaving several
lowresource languages unexplored in this direction. Sarcasm detection in low-resource Dravidian languages
like Tamil and Malayalam, are exacerbated by the complexity of linguistic structures, the rich variety
of expressions used in text to express sarcasm [6, 7] and the unavailability of annotated data, in these
languages. To address the challenges of detecting sarcastic content in code-mixed Tamil-English
and Malayalam-English text, in this paper, we, team MUCS describe the learning models submitted
to “Sarcasm Identification of Dravidian Languages in DravidianCodeMix@FIRE-2024" shared task 1
organized at Forum for Information Retrieval Evaluation (FIRE) 2024. Sarcasm detection is modeled as
a binary classification problem of assigning ’Non-sarcastic’ or ’Sarcastic’ label for the given code-mixed
Tamil-English and Malayalam-English text. Two distinct models: i) LSTM model trained using Keras
embeddings, and ii) mBERT+CNN model - a combination of TL (fine-tuning mBERT) and DL (CNN) for
building classifier, are proposed for sarcasm detection. As the datasets provided by the organizers for
this task are imbalanced, text augmentation techniques are used to increase the samples of the minority
class in the datasets. Text augmentation is expected to enhance the ability to identify and interpret
sarcastic comments efectively. The sample text with their corresponding labels from the given datasets
for code-mixed Tamil-English and Malayalam-English are shown in Table 1.</p>
      <p>The rest of the paper is organized as follows: Section 2 describes the recent literature on the sarcasm
identification in social media texts and Section 3 focuses on the description of the proposed models
followed by the experiments and results in Section 4. The paper concludes with future works in
Section 5.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>Sarcasm identification involves detecting ironic or mocking language, where the intended meaning is
often the opposite of the literal meaning of the words used. This task is challenging due to the nuanced
nature of sarcasm, which can vary greatly across diferent languages and contexts [ 8]. There have been
extensive research conducted on sarcasm detection in several Indian languages. Few of the relevant
works are described below:
1https://sites.google.com/view/dravidian-codemix-fire2024/</p>
      <p>For sarcasm identification in Tamil and Malayalam social media texts, Indirakanth et al. [9] proposed
Support Vector Machine (SVM) trained with Term Frequency-Inverse Document Frequency (TF-IDF) of
word n-grams in the range (1, 3) and TL (BERT, Distilled BERT (DistilBERT), Cross-lingual Language
Model - Robustly optimized BERT approach (XLM-RoBERTa)). Among their proposed models,
DistilBERT model outperformed other models with macro F1 scores of 0.68 and 0.63 for Tamil and Malayalam
datasets respectively. Pandey and Singh [10] implemented various learning models (Machine Learning
(ML), DL, and TL) for identifying sarcasm in the code-mixed Hindi-English text. The ML classifiers
(Naive Bayes (NB), Logistic Regression (LR), K-Nearest Neighbours (K-NN), SVM, Random Forest (RF),
Decision Tree (DT), Extreme Gradient Boosting (XGB)) were trained with TF-IDF of word uni-grams, DL
models (Deep Neural Network (DNN) CNN, LSTM) with Keras embeddings, and TL-based models using
ifne-tuned BERT. Further, they implemented a hybrid model by stacking LSTM network at the final layer
of BERT model. Among the proposed models, hybrid model outperformed other models with macro
F1 score of 0.98 for identifying sarcasm in the code-mixed Hindi-English text. Kolhe [11] developed
Marathi dataset with 2,400 Marathi tweets for binary classification, to identify sarcasm contents in
tweets and used various ML models (XGB, DT, RF and SVM) trained with TF-IDF of word unigrams, to
benchmark their dataset. Among the proposed ML models, XGBoost classifier trained with TF-IDF of
word unigrams, outperformed other models with a macro F1 score of 0.65.</p>
      <p>Bhaumik and Das [12] proposed TL-based models (mBERT and Multilingual Representations for
Indian Languages (MuRiL)) for sarcasm detection in Tamil-English and Malayalam-English dataset.
Among their proposed models MuRiL model obtained macro F1 scores of 0.781 and 0.731 for
TamilEnglish and Malayalam-English texts respectively. For sarcasm detection in Tamil and Malayalam
Dravidian code-mixed texts, Chanda et al. [13] implemented TL-based BERT model with an additional
layer of neural networks to classify comments as sarcastic or non-sarcastic and obtained macro F1
score of 0.72 for both languages. Eke et al. [14] experimented three approaches: i) a Bidirectional
LSTM (BiLSTM) model using Global Vectors (GloVe) embeddings, ii) a BERT-based model, and iii) a
feature fusion model incorporating BERT features, sentiment-related information, syntactic features,
and GloVe embeddings, for sarcasm detection in English social media and internet texts. Among their
proposed models BiLSTM and feature fusion models outperformed other models with macro F1 scores
of 0.98 and 0.80 on Twitter and Internet Argument Corpus version two (IAC-v2) datasets, respectively.
Kalaivani and Thenmozhi [15] implemented ML models (LR, RF, XGB, Support Vector Classifier (SVC),
and Gaussian Naive Binomial (GNB)) trained with Doc2Vec and TF-IDF of word unigrams, DL model
(Recurrent Neural Network with Long Short Term Memory (RNN-LSTM)) trained with with Keras
embeddings, and TL-based model (a pretrained BERT variants) fine-tuned with BERT features, for
identifying sarcasm in English text obtained from Twitter and Reddit forums. Among their proposed
models, TL-based BERT model obtained better F1 scores of 0.722 and 0.679 for the Twitter and Reddit
forums respectively.</p>
      <p>The related work illustrates that researchers have employed ML, DL, and TL models with various
features for sarcasm identification in Indian languages as well as in English. However, the performance
of not all models are promising. Further, the continuous evolution of user-generated content indicates
substantial opportunities for further research and innovation in this domain.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>The proposed methodology for sarcasm identification in code-mixed Tamil-English and
MalayalamEnglish texts includes: Text Augmentation, Pre-processing, Feature Extraction and Model Building. The
framework of the proposed methodology is shown in Figure 1 and the steps are explained below:</p>
      <sec id="sec-3-1">
        <title>3.1. Text Augmentation</title>
        <p>A significant variation in the number of instances belonging to the classes in a dataset indicates data
imbalance and this causes the learning models to be biased in favor of the majority class thereby
performing poorly on the minority class. This imbalance can be handled either by increasing samples
in the minority classes or decreasing samples in the majority classes. However, decreasing the samples
from majority class is not a good move if the data is less in size. This leads to increasing the size
of samples in minority class. The statistics of the code-mixed Tamil-English and Malayalam-English
datasets provided by the shared task organizers2 for sarcasm detection are shown in Table 2. From the
table, it can be observed that, in contrast to the ’Non-Sarcastic’ class, the number of samples belonging
to ’Sarcastic’ class in the Train sets of both the languages show an enormous diference and hence the
dataset is unbalanced.</p>
        <p>In order to balance the dataset, the number of samples can be expanded by either duplicating the
current samples or creating new synthetic data based on the current samples [16]. While duplicating the
existing samples just increase the size of the dataset, creating new synthetic samples add variation in the
data in addition to increasing the size of the dataset. The literature has suggested a number of methods
such as text augmentation, vector space resampling, and oversampling, to address the problem of class
imbalance by boosting the samples in the minority class. Text augmentation refers to a collection of
techniques used to artificially increase the size of the dataset [ 17]. This approach has become essential,
particularly for situations where the dataset exhibits an uneven distribution of classes or a relatively
low amount of labeled data. In this work, text augmentation techniques are used to boost the ’Sarcastic’
(minority) class samples in the Train sets of both the languages.</p>
        <p>Natural Language Processing Augmentation (NLPAug) - NLPAug3 is a Python library designed
to facilitate text augmentation for NLP tasks. It provides a variety of transformations for text data at
character-level, word-level, and sentence-level. NLPAug supports various techniques like synonym
replacement, random insertion and deletion of words and characters, and shufling of words and
characters [18]. It also includes more advanced techniques/augmenters integrated with pretrained
word embeddings (Word2Vec, GloVe, and BERT embeddings), contextual word embeddings, and back
translation, to generate new words/text. These techniques help to enhance the diversity and robustness
of the datasets, thereby improving the performance of the learning models. In this work, contextual word
embedding augmenter (NLPAugContextualWordEmbAug) technique is used to augment the ‘Sarcastic’
class samples by employing ‘insert’ option. This technique uses multilingualBERT4 - a pretrained
language model, to predict the words that have to be inserted at random positions in the original
2https://codalab.lisn.upsaclay.fr/competitions/19310
3https://nlpaug.readthedocs.io/en/latest/
4https://huggingface.co/google-bert/bert-base-multilingual-cased
sentence and the words are chosen automatically based on their contextual relevance by maintaining
the overall meaning and coherence of the sentence. In Tamil-English and Malayalam-English datasets,
sarcastic samples were increased to 21,740 and 10,689 samples respectively, to match the number
of non-sarcastic samples. The examples of Tamil-English and Malayalam-English sample texts and
augmented text using NLPAugContextualWordEmbAug technique are shown in Table 3.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Pre-processing</title>
        <p>Raw text often contains various forms of noise and irrelevant information that can adversely afect the
performance of text classification models. Pre-processing is an initial step in text classification that
involves transforming raw text data into a clean and structured format suitable for analysis. It also
helps to standardize and simplify the text, making it easier to extract meaningful patterns and features.
In this study, pre-processing includes removing user mentions (e.g., "@username"), numbers, URLs,
and HTML tags and converting emojis into their textual representations to maintain consistency and
enhancing the models ability to understand the context conveyed by emojis.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Model Building</title>
        <p>The augmented dataset is pre-processed and used to construct two distinct models: i) LSTM and ii)
mBERT+CNN. These models are chosen for their proven efectiveness in capturing contextual nuances
and handling sequential data, which are essential for tasks like sarcasm identification.</p>
        <sec id="sec-3-3-1">
          <title>3.3.1. LSTM - a Deep Learning Approach</title>
          <p>DL is a kind of ML that models and comprehends complicated patterns and representations in data
by using multi-layered Artificial Neural Networks (ANNs). The steps involved in implementing this
approach are given below:
• Text and Label Fusion - pre-processed augmented text data and the corresponding label is
fused to create a combined text-label format ensuring that both pieces of information are treated
as a unified input for the model during training. This enhances the model’s ability to learn the
relationship between text features and their respective labels, thereby improving the efectiveness
of the classification process. Incorporating label as part of textual content during training
allows the model to learn the nuanced relationship between textual features and their associated
sentiments. While labels are usually employed only as categorical identifiers in supervised
learning, integrating them with the text enables joint learning, enhancing the features captured
and refining overall classification performance.
• Feature Extraction - is a technique that involves converting raw text into numerical
representations that can be fed into ML algorithms and in this study, Keras embedding is employed
to facilitate this process. The Tokenizer class from Keras is utilized to convert text data into
sequences of integers based on the frequency of words, with a vocabulary size of 10,000 words
to maintain a reasonable model size, ensuring that the embeddings can efectively generalize to
unseen data without overfitting. These sequences are then padded to ensure uniform input length
for the model, specifically setting a maximum sequence length of 100. This step ensures that all
text inputs have the same dimensions, which is crucial for the model to process and learn from
the data efectively. Keras embeddings are language-independent and generate vectors based
on vocabulary, without capturing contextual nuances. This approach supports all languages,
including Tamil and Malayalam, by providing a uniform representation across diverse linguistic
contexts.
• Classifier Construction - LSTM is a type of RNN that is capable of learning and remembering
long term dependencies without encountering vanishing gradient descent or exploding problems
[19]. The model architecture has an embedding layer that maps each integer-encoded word
to a dense vector of 128 dimensions, capturing semantic relationships between words. This is
followed by an LSTM layer, which is designed to handle sequential data and capture long-term
dependencies within the text. Finally, a dense layer with a sigmoid activation function is used
to produce binary classification outputs. The model is compiled with the Adam optimizer and
binary cross-entropy loss function, suitable for binary classification tasks. It is then trained on
the padded training data and evaluated using validation/test data to monitor its performance.</p>
          <p>The use of LSTM enhances the model’s capacity to capture contextual dependencies, which is
beneficial for sarcasm identification. The hyperparameters and their values used in LSTM model is
shown in Table 4.</p>
        </sec>
        <sec id="sec-3-3-2">
          <title>3.3.2. mBERT+CNN model</title>
          <p>Integration of two or more techniques allows to utilize their complementary strengths for improved
classifier performance [ 20]. In this study, mBERT is used for feature extraction, and CNN is employed
for classification and the description of these are given below:
• Feature Extraction - uses TL for leveraging the knowledge learnt in solving a source task to
solve a target task, instead of building the model from scratch and the transformer models are
ifne-tuned for this purpose. mBERT is a transformer model that excels at capturing multilingual
and cross-lingual features and efectively manages and integrates mixed linguistic patterns and
semantic nuances through its pretrained language representations, allowing it to understand
and interpret the complex code-mixed language patterns found in Tamil-English and
MalayalamEnglish texts. These embeddings serve as rich feature representations, enabling the model to
benefit from pre-learned language structures and semantics [21].
• Classifier Construction - in recent years, ANNs have gained a lot of attention in designing
prediction models [22]. CNNs are a type of feed forward ANNs that applies convolutional filters to
learn and detect local patterns in data, when combined with pooling layers, efectively capturing
hierarchical features and structures for tasks such as text classification. CNN architecture is
employed for sarcasm detection, leveraging pretrained mBERT embeddings as features. It utilizes
mBERT embeddings having an vocabulary of size 110,000 and an embedding dimension of 768,
directly bypassing the need for an additional embedding layer, as these embeddings already
encapsulate rich text features. CNN applies convolutional layers with 100 filters and kernel sizes
of 2, 3, and 4 to capture local features from the mBERT embeddings. Each convolutional layer is
followed by a ReLU activation function and max pooling to reduce dimensionality. The pooled
feature maps are concatenated into a one-dimensional vector, which is then processed through a
dense layer with a softmax activation function to generate the final classification probabilities.
This approach combines the contextual understanding provided by mBERT with CNNs capability
to optimize the models performance in detecting sarcasm content in social media data.
The combination of mBERT for feature extraction and CNN for classification capitalizes on the strengths
of models, facilitating the processing of complex multilingual data while maintaining eficiency and
accuracy in sarcasm identification task. The hyperparameters and their values used in mBERT+CNN
model are shown in Table 5.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments and Results</title>
      <p>The datasets provided by the organizers of the shared task consists of YouTube comments for both
codemixed Tamil-English and Malayalam-English texts in native as well as Roman scripts, featuring a mix
of Tamil or Malayalam text with English. The performances of the proposed models with and without
augmentation on the Development and Test sets are shown in Tables 6 and 7 respectively. The results
in these tables reveal that the use of text augmentation has significantly improved the performances of
both the models. For Tamil-English Test set, LSTM models shows a notable increase in macro F1 scores
from 0.21 (without augmentation) to 0.45 (with augmentation) while mBERT+CNN models demonstrate
an enhanced performance with its macro F1 scores rising from 0.43 (without augmentation) to 0.74
(with augmentation). This indicates that text augmentation aids in enhancing model performance for
sarcasm detection. For Malayalam-English Test set, while the macro F1 scores of LSTM models have
increased from 0.15 (without augmentation) to 0.19 (with augmentation), mBERT+CNN models has
maintained a consistent macro F1 score of 0.72 (for both with and without augmentation), indicating
that augmentation had minimal efect on their performances. Further, from Table 7 it is clear that
mBERT+CNN models obtained better macro F1 scores of 0.74 and 0.72 for code-mixed Tamil-English
and Malayalam-English texts respectively. The comparison of macro F1 scores of all the participating
teams for the sarcasm detection task in both code-mixed Tamil-English and Malayalam-English text are
shown in Figure 2 (a) and 2 (b) respectively.</p>
      <p>Few misclassified Malayalam-English samples using mBERT+CNN model along with the actual and
predicted labels and the probable reasons for misclassification are shown in Table 8.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion and Future Work</title>
      <p>In this paper, we - team MUCS, describe the models submitted to ’Sarcasm Identification of Dravidian
Languages in DravidianCodeMix@FIRE-2024’ a shared task organized at ’FIRE 2024’, to distinguish
between ’Non-sarcastic’ and ’Sarcastic’ comments in code-mixed Tamil-English and Malayalam-English.
Experiments are carried out with two proposed models: i) LSTM model - trained using Keras embeddings
and ii) mBERT+CNN model - a combination of TL (fine-tuning mBERT model) and DL approach (CNN)
for building classifier, for sarcasm identification. Further, to overcome the data imbalance issue in the
dataset, text augmentation technique is explored using Contextual word embedding augmenter from
NLPAug library, to increase the number of samples in the minority class. Among the proposed models,
mBERT+CNN model outperformed LSTM model with macro F1 scores of 0.74 and 0.72 for Tamil-English
and Malayalam-English datasets securing 1st and 4th ranks respectively. Advanced text augmentation
techniques followed by eficient text representation methods and context-aware models to enhance
sarcasm detection in social media comments will be explored further.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used ChatGPT in order to: Grammar and spelling
check. After using this tool/service, the author(s) reviewed and edited the content as needed and take(s)
full responsibility for the publication’s content.
[5] R. Pandey, A. Kumar, J. P. Singh, S. Tripathi, A Hybrid Convolutional Neural Network for Sarcasm
Detection from Multilingual Social Media Posts, in: Multimedia Tools and Applications, Springer,
2024, pp. 1–29.
[6] B. R. Chakravarthi, N. Sripriya, B. Bharathi, K. Nandhini, S. C. Navaneethakrishnan, T. Durairaj,
R. Ponnusamy, P. K. Kumaresan, K. K. Ponnusamy, C. Rajkumar, Overview of the Shared Task on
Sarcasm Identification of Dravidian Languages (malayalam and tamil) in dravidiancodemix, in:
Forum of Information Retrieval and Evaluation FIRE-2023, 2023.
[7] N. Sripriya, T. Durairaj, K. Nandhini, B. Bharathi, K. K. Ponnusamy, C. Rajkumar, P. K. Kumaresan,
R. Ponnusamy, C. Subalalitha, B. R. Chakravarthi, Findings of Shared Task on Sarcasm Identification
in Code-Mixed Dravidian Languages, in: FIRE 2023, volume 16, 2023, p. 22.
[8] R. González-Ibánez, S. Muresan, N. Wacholder, Identifying Sarcasm in Twitter: a Closer Look, in:
Proceedings of the 49th annual meeting of the association for computational linguistics: human
language technologies, 2011, pp. 581–586.
[9] V. Indirakanth, D. Udayakumar, T. Durairaj, B. Bharathi, Sarcasm Identification of Dravidian</p>
      <p>Languages (Malayalam and Tamil)., in: FIRE (Working Notes), 2023, pp. 327–335.
[10] R. Pandey, J. P. Singh, BERT-LSTM Model for Sarcasm Detection in Code-mixed Social Media Post,
in: Journal of Intelligent Information Systems, volume 60, Springer, 2023, pp. 235–254.
[11] P. K. P. P. S. Kolhe, MarathiSarc: A Marathi Tweets Dataset for Automatic Sarcasm Detection of</p>
      <p>Marathi Tweets, 2022.
[12] A. B. Bhaumik, M. Das, Sarcasm Detection in Dravidian Code-Mixed Text Using Transformer-Based</p>
      <p>Models., in: FIRE (Working Notes), 2023, pp. 249–258.
[13] S. Chanda, A. Mishra, S. Pal, Sarcasm Detection in Tamil and Malayalam Dravidian Code-Mixed</p>
      <p>Text., in: FIRE (Working Notes), 2023, pp. 336–343.
[14] C. I. Eke, A. A. Norman, L. Shuib, Context-based Feature Technique for Sarcasm Identification
in Benchmark Datasets using Deep Learning and BERT Model, in: IEEE Access, volume 9, IEEE,
2021, pp. 48501–48518.
[15] A. Kalaivani, D. Thenmozhi, Sarcasm Identification and Detection in Conversion Context using
BERT, in: Proceedings of the Second Workshop on Figurative Language Processing, 2020, pp.
72–76.
[16] L. Wang, M. Han, X. Li, N. Zhang, H. Cheng, Review of Classification Methods on Unbalanced</p>
      <p>Data sets, in: Ieee Access, volume 9, IEEE, 2021, pp. 64606–64628.
[17] C. Shorten, T. M. Khoshgoftaar, B. Furht, Text Data Augmentation for Deep Learning, in: Journal
of big Data, volume 8, Springer, 2021, p. 101.
[18] P. Gujjar, P. K. HR, Natural Language Processing using Text Augmentation for Chatbot, in: 2022
International Conference on Artificial Intelligence and Data Engineering (AIDE), IEEE, 2022, pp.
248–251.
[19] S. Hochreiter, Long Short-term Memory, in: Neural Computation MIT-Press, 1997.
[20] J. Sansana, M. N. Joswiak, I. Castillo, Z. Wang, R. Rendall, L. H. Chiang, M. S. Reis, Recent Trends
on Hybrid Modeling for Industry 4.0, in: Computers &amp; Chemical Engineering, volume 151, Elsevier,
2021, p. 107365.
[21] J. Devlin, M. Chang, K. Lee, K. Toutanova, BERT: Pre-training of Deep Bidirectional Transformers
for Language Understanding, in: CoRR, volume abs/1810.04805, 2018. URL: http://arxiv.org/abs/
1810.04805. arXiv:1810.04805.
[22] P. Mehndiratta, S. Sachdeva, D. Soni, Detection of Sarcasm in Text Data using Deep Convolutional
Neural Networks, in: Scalable Computing: Practice and Experience, volume 18, 2017, pp. 219–228.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. N</given-names>
            , B. B,
            <surname>N. K</surname>
          </string-name>
          , T. Durairaj,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Kumaresan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Rajkumar, Overview of Sarcasm Identification of Dravidian Languages in DravidianCodeMix@FIRE-2024, in: Forum of Information Retrieval and Evaluation FIRE -</article-title>
          <year>2024</year>
          , DAIICT , Gandhinagar,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <article-title>Hope Speech Detection in YouTube Comments, in: Social Network Analysis and Mining</article-title>
          , volume
          <volume>12</volume>
          , Springer,
          <year>2022</year>
          , p.
          <fpage>75</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hande</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Kumaresan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Priyadharshini</surname>
          </string-name>
          ,
          <article-title>How Can We Detect Homophobia and Transphobia? Experiments in a Multilingual Code-mixed Setting for Social Media Governance</article-title>
          , in:
          <source>International Journal of Information Management Data Insights</source>
          , volume
          <volume>2</volume>
          ,
          <string-name>
            <surname>Elsevier</surname>
          </string-name>
          ,
          <year>2022</year>
          , p.
          <fpage>100119</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Sripriya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bharathi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Nandhini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. Chinnaudayar</given-names>
            <surname>Navaneethakrishnan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Durairaj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Kumaresan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Rajkumar, Overview of the Shared Task on Sarcasm Identification of Dravidian Languages (Malayalam and Tamil) in DravidianCodeMix, in: Forum of Information Retrieval and Evaluation FIRE -</article-title>
          <year>2023</year>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>