<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Investigation of Machine Learning and Transformer Models for Sarcasm Detection in Dravidian Languages</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Malliga Subramanian</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Ananthakumar S</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Deepiga P</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Dharshini S</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kogilavani S V</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Computer Science and Engineering</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Kongu Engineering College Erode Tamil Nadu</institution>
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Sarcasm detection in natural language is a critical task for sentiment analysis, particularly in resource-constrained languages like Malayalam, a Dravidian language. We secured 6th place in the Sarcasm Identification of Dravidian Languages (Malayalam) track at DravidianCodeMix@FIRE-2024. The complexity of sarcasm, which often relies on context, tone, and cultural understanding, poses unique challenges for machine learning models. This study investigates the efectiveness of various machine learning and deep learning techniques in identifying sarcasm in Malayalam text. We employ a diverse set of models, including RoBERTa, Convolutional Neural Networks (CNN), Multilayer Perceptron (MLP), Gated Recurrent Units (GRU), Recurrent Neural Networks (RNN), Random Forests (RF), Hidden Markov Models (HMM), Logistic Regression (LR), K-Nearest Neighbors (KNN), and Gaussian Mixture Models (GMM).</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Sarcasm, characterized by the use of irony to mock or convey contempt, is one of the most complex
linguistic phenomena to detect in natural language processing (NLP) [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. In human communication,
sarcasm often relies on nuanced cues such as tone, facial expressions, or context that are dificult to
interpret in written text. Its proper identification is crucial for sentiment analysis, opinion mining,
and emotion recognition [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. The challenge of sarcasm detection becomes even more pronounced in
less-resourced languages, where annotated datasets and linguistic resources are limited. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] Malayalam,
a Dravidian language spoken primarily in the Indian state of Kerala, presents unique dificulties for
sarcasm detection due to its rich morphology, complex grammatical structure, and diverse expressions.
[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] In recent years, several machine learning and deep learning techniques have been employed to
address sarcasm detection [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], but their application in Malayalam is relatively underexplored. Previous
studies have predominantly focused on widely spoken languages like English, leaving a significant gap
in NLP research for regional languages like Malayalam. The limited availability of annotated sarcasm
datasets in Malayalam further exacerbates this problem, making it dificult to train robust models for
sarcasm identification. In this study, we focus on sarcasm identification in Malayalam using a variety
of machine learning and deep learning models. [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] provides a comprehensive overview of sarcasm
detection techniques across various languages, including applications for Dravidian languages.We
explore transformer-based model like RoBERTa, which have demonstrated state-of-the-art performance
in NLP tasks by leveraging contextual embeddings. Additionally, we apply deep learning architectures
such as Convolutional Neural Networks (CNN), Multilayer Perceptron (MLP), Gated Recurrent Units
(GRU), and Recurrent Neural Networks (RNN), which excel in capturing semantic and sequential
information from text. To benchmark the performance of traditional machine learning approaches,
we also employ models like Random Forests (RF), Hidden Markov Models (HMM), Logistic Regression
(LR), K-Nearest Neighbors (KNN), and Gaussian Mixture Models (GMM). This study aims to evaluate
the efectiveness of these models in detecting sarcasm in Malayalam, identifying key challenges, and
comparing the performance of diferent approaches. By exploring a diverse range of models, we seek to
provide insights into the strengths and limitations of machine learning and deep learning techniques for
sarcasm detection in low-resource languages like Malayalam, ultimately contributing to the development
of more robust NLP tools for Dravidian languages.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Literature Survey</title>
      <p>
        Sarcasm detection is a crucial area of research in natural language processing (NLP), especially for
sentiment analysis, opinion mining, and emotion recognition. The survey [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] provides a comprehensive
overview of sarcasm detection techniques across various languages, including applications for Dravidian
languages.The survey [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] explores sarcasm detection using machine learning techniques applicable to
microblogs and social media data. The survey [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] findings of the shared task on ofensive language
identification in Dravidian languages. Proceedings of the First Workshop on Speech and Language
Technologies for Dravidian Languages. This shared task includes sarcasm detection in Malayalam.
While substantial progress has been made in sarcasm ientification for English and other well-resourced
languages, the research for low-resource languages, including Dravidian languages such as
Malayalam, remains limited.The survey [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] Proceedings of the 2017 Conference on Empirical Methods in
Natural Language Processing (EMNLP). The survey [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] gives proceedings of the Second Workshop on
Multilinguality and Code Switching (MultiLingCode). Discussdes sarcasm detection in Tamil, which
shares linguistic similarities with Malayalam.The survey [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] building a dictionary of afective words
for a sentiment analysis of online messages. Journal of Language Resources and Evaluation, 50(3),
665-691. This paper focuses on sentiment analysis, an underlying concept for sarcasm detection in
Malayalam.This paper discusses the role of personalized contexts in detecting sarcasm, useful for
Dravidian languages.the survey [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] gives proceedings of the Third Workshop on Computational
Approaches to Linguistic Code-Switching. This work explores the classification of code-mixed Dravidian
languages, applicable to sarcasm detection in Malayalam.The complexity of sarcasm detection arises
from its reliance on context, tone, and cultural nuances, which make it a non-trivial task for machine
learning and deep learning models. The survey [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] Proceedings of the International Conference on
Emerging Technologies in Computer Engineering: Machine Learning and Internet of Things. Discusses
the role of sarcasm in sentiment analysis, relevant to detecting sarcasm in Malayalam. The survey [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]
presents machine learning techniques for sarcasm detection, useful for Dravidian languages. The survey
[
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] gives proceedings of the First Workshop on Speech and Language Technologies for Dravidian
Languages. Ofers insights into sentiment and sarcasm detection in code-mixed Indian languages,
including Malayalam.The survey [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] show their findings highlight the efectiveness of diferent modeling
approaches in understanding sarcasm in textual data. The survey [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] provides an overview of the
models submitted by participants for the task of sarcasm identification in Dravidian languages, as
presented in DravidianCodeMix@FIRE-2024. It highlights the methodologies employed, the diversity of
approaches, and the overall contributions of each submission, aiming to enhance understanding and
improve future eforts in this area of research.
      </p>
      <sec id="sec-2-1">
        <title>2.1. Sarcasm Detection in English and Other Major Languages</title>
        <p>
          The earliest approaches to sarcasm detection relied on rule-based and lexical analysis techniques,
which were primarily centered on identifying specific syntactic or semantic patterns. Davidov et al.
(2010) introduced a semi-supervised sarcasm detection technique that leveraged features like patterns,
punctuation, and n-grams from Twitter data in English [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ]. These methods, though efective in
specific scenarios, struggled with generalization due to the nuanced nature of sarcasm, which is highly
context-dependent. As machine learning models advanced, supervised learning methods such as
Support Vector Machine (SVM), Random Forests (RF), and Logistic Regression (LR) became more widely
used in sarcasm detection tasks [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ]. These models often relied on hand-crafted features, including
ngrams, sentiment lexicons, and part-of-speech tags. While these methods provided improvements, their
reliance on manually engineered features limited their ability to capture deeper contextual and linguistic
nuances essential for sarcasm detection. More recently, transformers like BERT (Bidirectional Encoder
Representations from Transformers) and its variants (e.g., RoBERTa, ALBERT) have set new benchmarks
in NLP tasks by utilizing self-attention mechanisms to understand complex word relationships in context.
These models have achieved state-of-the-art performance in sarcasm detection for English due to their
ability to leverage contextual embeddings, which capture intricate meanings within sentences.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Sarcasm Detection in Dravidian Languages</title>
        <p>
          Despite the advances in English sarcasm detection, research in sarcasm identification for Dravidian
languages like Malayalam has been sparse. Dravidian languages are morphologically rich and exhibit
complex syntactic structures, making sarcasm detection particularly challenging [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]. The lack of large
annotated datasets for these languages adds to the dificulty of building reliable models. Most research
in Malayalam has focused on sentiment analysis and emotion detection, often using shallow machine
learning models like Naive Bayes, SVM, and RF. For instance, Kumar et al. (2020) explored sentiment
analysis in Tamil and Malayalam using traditional machine learning techniques [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ], demonstrating that
these languages require more advanced methods for sarcasm detection. The reliance on hand-crafted
features limits the efectiveness of these models in capturing the full complexity of sarcasm.
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Deep Learning and Transformer Models for Malayalam</title>
        <p>
          Given the success of deep learning models in other languages, there is growing interest in applying
these methods to sarcasm detection for Malayalam. Models like CNN, Multilayer Perceptron (MLP),
and Gated Recurrent Units (GRU) have shown promise in handling the intricacies of text classification
tasks by capturing semantic and sequential information in the data. [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] However, the application of
these models to Malayalam sarcasm detection is still underexplored due to the limited availability of
labeled data. Transformer-based models such as RoBERTa, and GPT have revolutionized NLP tasks by
capturing complex relationships in the text through self-attention mechanisms [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. These models could
significantly improve sarcasm detection in Malayalam by leveraging transfer learning, where models
pre-trained on large datasets (such as multilingual BERT) are fine-tuned for Malayalam-specific tasks.
Preliminary research in sentiment analysis for Dravidian languages using transformer models suggests
that they have great potential for sarcasm detection as well, provided there are enough annotated
resources for fine-tuning.
        </p>
      </sec>
      <sec id="sec-2-4">
        <title>2.4. Challenges and Future Directions</title>
        <p>One of the primary challenges in sarcasm detection for Malayalam is the lack of large-scale annotated
datasets. Manual annotation of sarcasm is resource-intensive due to the nuanced nature of the task.
Additionally, Malayalam exhibits rich morphology, including inflectional changes that make text
normalization and feature extraction challenging. Future research should focus on building more
extensive annotated corpora and developing hybrid approaches that combine traditional machine
learning models with deep learning architectures to better capture the complexity of sarcasm. In
conclusion, while sarcasm detection for Malayalam and other Dravidian languages is in its early stages,
there is significant potential to improve the task by adopting modern deep learning techniques and
transfer learning approaches.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Materials and Methods</title>
      <sec id="sec-3-1">
        <title>3.1. Dataset Description</title>
        <p>The dataset for sarcasm identification in Malayalam consists of social media posts and online comments,
annotated to classify each text as either sarcastic or non-sarcastic. The texts are sourced from platforms
like Twitter and Facebook, where sarcasm is commonly used. Each post is labeled based on its contextual
meaning and linguistic cues. Sarcastic texts (labeled as 1) convey an ironic or contradictory meaning
compared to their literal interpretation, while non-sarcastic texts (labeled as 0) reflect straightforward
expressions. The dataset is preprocessed to remove special characters, URLs, and stop-words, and it
may include features like punctuation and polarity to help models better understand the underlying
sarcasm. Models trained on this dataset are evaluated on metrics such as accuracy, precision, recall, and
F1-score to assess their performance in distinguishing between sarcastic and non-sarcastic Malayalam
texts.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Pre-processing</title>
        <p>Pre-processing for sarcasm identification in Malayalam involves a series of steps designed to clean
and standardize the text, preparing it for meaningful analysis. Given the language’s rich morphology
and frequent code-switching with English, the first step is data cleaning, which involves removing
special characters, URLs, and irrelevant stop-words that do not contribute to the sarcastic content.
This is particularly important when handling mixed-language text, where Malayalam and English
are interspersed. Following data cleaning, text normalization is performed to standardize the words.
Tokenization is the process of breaking down the text into smaller units, such as words or sub-words,
which is crucial for understanding Malayalam’s script and grammar. Stemming and lemmatization
are applied to reduce words to their root forms, which helps in managing the language’s complex
inflectional morphology. For example, diferent forms of the same word can be converted to a common
root, making it easier for models to understand the core meaning. Punctuation, which plays a significant
role in sarcasm detection, is deliberately retained during pre-processing. Elements like exclamation
marks, ellipses, or question marks can alter the tone of a statement and often indicate sarcasm, so they
are treated as features rather than noise. Additionally, handling agglutination is crucial in Malayalam,
as the language frequently combines multiple morphemes into a single word, creating compound words.
By breaking these down into individual morphemes or identifying the root forms and afixes, the
system can better interpret the word’s meaning in context. These comprehensive pre-processing steps
ensure the data is clean, normalized, and well-structured, laying the groundwork for efective sarcasm
identification.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Feature Extraction</title>
        <p>Feature extraction for sarcasm identification in Malayalam involves techniques aimed at capturing
linguistic patterns and contextual nuances that distinguish sarcastic expressions. The process begins
with n-gram features, such as unigrams, bigrams, and trigrams, which help identify word sequences
that may indicate sarcasm when analyzed in context, for instance, phrases where the literal meaning
contrasts with the intended tone. TF-IDF (Term Frequency-Inverse Document Frequency) is another
important approach, assigning weights to words based on their frequency and uniqueness, highlighting
terms that might carry ironic or sentimentally charged meanings. Word embeddings, like Word2Vec
and BERT, enhance the understanding of sarcasm by providing dense vector representations of words,
with Word2Vec capturing semantic similarity and BERT ofering contextual embeddings that consider
surrounding words. Sentiment and polarity features further contribute by detecting contrasts between
the expressed sentiment and actual intent, where a positive word used in a negative scenario might
signal sarcasm. Part-of-speech (POS) tagging adds a syntactic dimension, helping to identify sarcastic
patterns through the use of specific adjectives or unexpected word combinations. Given Malayalam’s
morphological richness, handling agglutination by decomposing complex words into their base forms
and grammatical components is also essential. Additionally, recognizing idiomatic phrases and
punctuation marks, such as exclamation points or ellipses, plays a significant role, as these elements often carry
ironic undertones. Together, these techniques allow the model to capture both the linguistic features
and contextual cues necessary for accurately detecting sarcasm in the text.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Proposed Classifiers</title>
        <p>For sarcasm identification in Malayalam, performance metrics are used to evaluate how well the model
distinguishes between sarcastic and non-sarcastic text. Accuracy measures the percentage of correctly
classified examples but may not fully reflect performance when sarcasm is rare. Precision evaluates how
many of the texts predicted as sarcastic are truly sarcastic, while Recall measures the model’s ability
to identify actual sarcastic texts, minimizing missed cases (false negatives). The F1-Score combines
precision and recall into a balanced metric, useful when sarcasm is less frequent. A confusion matrix
provides a detailed view of the model’s errors, showing true positives, false positives, true negatives, and
false negatives. These metrics together ofer a comprehensive understanding of the model’s strengths
and weaknesses.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Dataset</title>
      <p>The Training Dataset consists of 13,188 instances, including 2,499 categorized as sarcastic and 10,689
as non-sarcastic. In contrast, the Test Dataset includes 2,826 instances, with 521 labeled as sarcastic
and 2,305 as non-sarcastic. Together, these datasets provide a robust foundation for training models to
detect sarcasm in Malayalam text. The significant number of non-sarcastic examples in both datasets is
crucial for efectively evaluating the model’s performance during testing.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion and Future Work</title>
      <p>Sarcasm identification in Malayalam, a Dravidian language, poses challenges due to its rich morphology,
syntactic complexity, and frequent code-switching with English. Utilizing advanced classifiers like
RoBERTa, CNNs, and GRUs alongside traditional models such as Random Forests and Logistic Regression,
substantial progress has been made in detecting sarcasm, with models efectively capturing linguistic
cues and contextual nuances. Pre-processing techniques like tokenization, sentiment analysis, and word
embeddings have contributed to improved performance. However, further optimization is required
to fully capture the complexity of sarcastic expressions in Malayalam. Future work could focus on
ifne-tuning advanced language models for Malayalam, leveraging cross-lingual models from other
Dravidian languages, and building larger, diverse datasets from social media, including code-mixed
content. Multimodal approaches incorporating text with audio or visual cues could further enhance
sarcasm detection. Additionally, incorporating cultural and idiomatic knowledge into the models would
provide a deeper understanding of sarcasm specific to Malayalam, enabling more accurate predictions.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>The author(s) have not employed any Generative AI tools.</p>
    </sec>
    <sec id="sec-7">
      <title>6. References</title>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>Joshi</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bhattacharyya</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Carman</surname>
            ,
            <given-names>M. J.</given-names>
          </string-name>
          (
          <year>2017</year>
          ).
          <article-title>Automatic Sarcasm Detection: A Survey</article-title>
          .
          <source>ACM Computing Surveys (CSUR)</source>
          ,
          <volume>50</volume>
          (
          <issue>5</issue>
          ),
          <fpage>1</fpage>
          -
          <lpage>22</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <surname>Mukherjee</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          , and Bala,
          <string-name>
            <surname>P. K.</surname>
          </string-name>
          (
          <year>2017</year>
          ).
          <article-title>Sarcasm detection in microblogs using Naïve Bayes and fuzzy clustering</article-title>
          .
          <source>Journal of Computer Science</source>
          ,
          <volume>13</volume>
          (
          <issue>5</issue>
          ),
          <fpage>141</fpage>
          -
          <lpage>150</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>Chakravarthi</surname>
            ,
            <given-names>B. R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Muralidaran</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Priyadharshini</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <surname>Ghosh</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Veale</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          (
          <year>2017</year>
          ).
          <article-title>Magnets for Sarcasm: Making Sarcasm Detection Timely</article-title>
          , Contextual, and Very Personal.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <surname>Annamalai</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Muralidaran</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          (
          <year>2021</year>
          ).
          <article-title>Cross-lingual Sarcasm Detection for Code-Mixed Tamil-English Text</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <surname>Cruz</surname>
            ,
            <given-names>F. L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Troyano</surname>
            ,
            <given-names>J. A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pontes</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Vallejo</surname>
            ,
            <given-names>C. G.</given-names>
          </string-name>
          (
          <year>2016</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>Chakravarthi</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Priyadharshini</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kumar</surname>
            ,
            <given-names>M. A.</given-names>
          </string-name>
          , and Jose, N. (
          <year>2022</year>
          ).
          <article-title>Dravidian Code-Mixed Text Classification using Transformer-based Models</article-title>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <surname>Kumar</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Garg</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          (
          <year>2019</year>
          ).
          <source>Sarcasm Detection and Its Impact on Sentiment Analysis: A Review.</source>
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <surname>Swain</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Rajendran</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Singh</surname>
            ,
            <given-names>A. K.</given-names>
          </string-name>
          (
          <year>2017</year>
          ).
          <article-title>A comparison of machine learning techniques for sarcasm detection</article-title>
          .
          <source>Journal of Computer Science and Technology</source>
          ,
          <volume>32</volume>
          (
          <issue>4</issue>
          ),
          <fpage>729</fpage>
          -
          <lpage>744</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Chakravarthi</surname>
            ,
            <given-names>B. R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kumar</surname>
            ,
            <given-names>S. A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Priyadharshini</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          , and
          <string-name>
            <surname>Muralidaran</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          (
          <year>2020</year>
          ).
          <article-title>Sentiment analysis of code-mixed Indian languages: An overview of SAIL Code-mixed shared task at ICON-20.</article-title>
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <surname>Kumar</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Garg</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          (
          <year>2019</year>
          ).
          <article-title>Empirical study of shallow and deep learning models for sarcasm detection using context in benchmark datasets</article-title>
          .
          <source>Journal of ambient intelligence and humanized computing</source>
          ,
          <fpage>1</fpage>
          -
          <lpage>16</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <surname>Chakravarthi</surname>
            ,
            <given-names>B. R.</given-names>
          </string-name>
          , N, S.,
          <string-name>
            <surname>B</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            , K, N.,
            <surname>Durairaj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            ,
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            ,
            <surname>Kumaresan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            ,
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            ,
            <surname>Rajkumar</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          (
          <year>2024</year>
          ).
          <article-title>Overview of sarcasm identification of Dravidian languages in DravidianCodeMix@FIRE-2024</article-title>
          .
          <article-title>In Forum of Information Retrieval and Evaluation FIRE - 2024</article-title>
          . DAIICT, Gandhinagar.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>