<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Detecting Sarcasm in Social Media Text Using Indic Transliteration and Machine Learning Techniques</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Kogilavani S V</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Malliga Subramnanian</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Prenitha S P</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Varshini S H</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Arunachalam M</string-name>
        </contrib>
      </contrib-group>
      <abstract>
        <p>Sarcasm, which is frequently characterized by irony or ridicule, is a complicated phenomena in language when the intended meaning difers from the literal utterance. Sarcasm detection in text poses major hurdles to natural language processing (NLP), especially in social media situations. In contrast to conventional sentiment analysis, sarcasm frequently hides its true meaning beneath colloquial language, making it challenging to identify without taking into account the context and minute changes in meaning.Using transliterated datasets, this study focuses on sarcasm recognition in Indic languages, particularly Tamil and Malayalam. We improve the text's accessibility for machine learning models by transliterating the native scripts into standard forms. The eficacy of four models in identifying sarcasm was tested: Decision Tree (DT), K-Nearest Neighbors (KNN), Support Vector Machine (SVM), and Logistic Regression (LR). With an F1-score of 0.3710 and accuracy rates of 83.62% for Malayalam and 75.50% for Tamil, Logistic Regression was the model that performed the best overall.Our system demonstrated the resilience of our method by achieving a noteworthy rank 9 in a competitive sarcasm detection assignment.These ifndings highlight how crucial it is to combine customized machine learning models with transliteration in order to identify sarcasm in regional languages.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Sarcasm detection</kwd>
        <kwd>Transliteration</kwd>
        <kwd>Natural Language Processing (NLP)</kwd>
        <kwd>Tamil</kwd>
        <kwd>Malayalam</kwd>
        <kwd>Code-mixed languages</kwd>
        <kwd>Machine Learning</kwd>
        <kwd>Logistic Regression</kwd>
        <kwd>K-Nearest Neighbors (KNN)</kwd>
        <kwd>Support Vector Machine (SVM)</kwd>
        <kwd>Decision Tree (DT)</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The practice of translating a word or phrase from one writing system to another while preserving the
original writing system’s sounds or letters is known as transliteration.When used skillfully,
transliterations and code-switching in languages such as Tamil and Malayalam can yield insightful linguistic
information that enhances the ability to recognize sarcasm [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Although these components appear to
add more labour to the task at hand, they actually create a more realistic representation of language by
combining phonetic and cultural elements with local linguistic patterns using English script.
Transliterations reveal contextual details that would otherwise be lost in translation since they accurately capture
the sound of words in their native language.By incorporating advanced machine learning techniques,
transliterations become a benefit rather than a disadvantage. Transliterations can be recognized and
interpreted by machine learning algorithms, especially those built to handle multilingual data. This
successfully bridges the gap between the original language and its English-script counterpart. These
algorithms are able to identify patterns that disclose the speaker’s intention, even when it is obscured
by sarcastic tones, by examining both the transliterated text and its surface-level content. This is made
possible by the fact that, provided phonetic consistencies are preserved, feature extraction models—like
those trained on English data—can nevertheless extract semantic information from transliterated text.
Text that has been transliterated into the Roman script adheres to a systematic, phonetically consistent
sound mapping. Therefore, despite the script variation, feature extraction algorithms use these phonetic
signals to identify well-known patterns, including sentiment changes or tonal nuances. Models are
taught to identify characteristics such as contextual dependencies, n-grams, and word embeddings in
transliteration cases, which preserve important linguistic elements of the source language. Furthermore,
Both native and transliterated data are used to train embedding methods such as Word2Vec, which
capture both surface-level features and more profound semantic implications. remove a line from this
and it should stil make sense
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Literature Review</title>
      <p>
        Sarcasm recognition in Indian languages has been a major focus of research. The study [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] on detecting
sarcasm in Hindi used language-specific methods such as transliteration and tokenisation from the
Indic NLP toolbox. Sarcasm detection
      </p>
      <p>Language nuances and code switching cause problems in Hindi, as they do in other Indian languages.
The study discovered that collecting features from Indic scripts using models like SVM and Random
Forest can improve sarcasm detection by leveraging language-specific characteristics.</p>
      <p>
        The author at [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] investigated a multimodal approach to sarcasm detection, using textual and visual
data. This study discovered that sarcasm is frequently expressed through imagery, emoticons, and social
media replies. By combining text-based characteristics with visual sentiment signals, their technique
outperformed purely text-based models in accuracy. The use of multimodal approaches allows for a
more full investigation of sarcasm, as it captures context that text alone may overlook.
      </p>
      <p>
        Both the studies by [
        <xref ref-type="bibr" rid="ref4 ref5">4, 5</xref>
        ] used machine learning models including Decision Trees, SVM, and Neural
Networks to identify sarcasm. The author used these models to anticipate sarcastic intent by analysing
the text’s syntactic and semantic components. In [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ], the author applied neural networks to
identify sarcasm across code-mixed languages, highlighting the significance of deep learning in dificult
multilingual contexts.
      </p>
      <p>
        The author [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] emphasised the role of transliteration in handling Indian scripts such as Tamil.
Transliteration bridges the gap between native scripts and their Romanised counterparts, which is
required for sarcasm detection in code-mixed datasets. Accurate transliteration not only protects
phonetic integrity, but it also improves feature extraction for machine learning models by retaining
important linguistic nuances. This, in turn, enhances the model’s capacity to recognise sophisticated
language elements like sentiment, intent, and sarcasm in multilingual settings. Transliteration also
increases performance on NLP tasks such as named entity recognition (NER), tokenisation, and
crosslingual transfer learning by providing a consistent representation of words across languages and scripts.
Transliteration has been shown in studies to reduce text ambiguity and improve overall classification
accuracy, especially in code-mixed circumstances where Romanised and local characters are merged.
      </p>
      <p>
        Finally, the author of [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] emphasised the role of transliteration in managing Indic scripts like Tamil.
Transliteration bridges the gap between native scripts and their Romanised counterparts, which is
required for sarcasm detection in code-mixed datasets. Their findings demonstrate that correct
transliteration preserves phonetic integrity, allowing for superior feature extraction in machine learning
models.
      </p>
      <p>These researches emphasise the significance of dealing with linguistic diversity, exploiting multimodal
inputs, and using advanced machine learning algorithms. Combining these approaches considerably
increases model robustness and accuracy in identifying sarcasm in Indian languages.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Dataset Description</title>
      <p>The dataset used in this research features mixed-language utterances that include Tamil-English and
Malayalam-English combinations, categorized as either sarcastic or non-sarcastic. The dataset is
available at: https://codalab.lisn.upsaclay.fr/competitions/19310#participate. Text samples in Malayalam
and Tamil that have been separated into training, validation, and test subsets make up the sarcasm
recognition dataset. There are 21,740 non-sarcastic and 7,830 sarcastic text in the training set and 4,630
non-sarcastic and 1,706 sarcastic text in the validation set for the Tamil dataset. 6,338 text samples make
up the test set, which will be used for label-free evaluation. There are 10,689 non-sarcastic and 2,499
sarcastic text in the training set and 2,305 non-sarcastic and 521 sarcastic text in the validation set of
the Malayalam dataset. The model’s performance on unseen data will also be evaluated using the 2,826
unlabeled text samples in the test set. This dataset is set up to train and validate models for sarcasm
detection in both languages.The dataset is an extensive collection for sarcasm identification in Tamil
and Malayalam since it spans a wide range of situations and domains, from oficial writing to social
media posts. The model’s capacity to generalize well to new data is improved by the quantity of the
dataset, which guarantees that it is trained on a variety of samples. Furthermore, a diferent validation
set makes sure that the models are adjusted to prevent overfitting, which strengthens their resilience.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Methodology</title>
      <sec id="sec-4-1">
        <title>4.1. Dataset Preprocessing</title>
        <p>A thorough preparation pipeline is crucial when working with both native-script and Romanized
Tamil text data. The first step is transliterating Romanized Tamil into native Tamil script using the
indic-transliteration module to ensure a consistent language representation during feature extraction.</p>
        <p>Following transliteration, preprocessing is carried out to clean and standardize the data. All text is
converted to lowercase to maintain uniformity, and punctuation, special characters, and non-alphabetical
symbols are removed, as they don’t contribute to the text’s semantic meaning. Stopwords, common
words like "and" ,"the" and "is" which don’t add significant meaning, are also removed using a
specially compiled Tamil stopwords list, expanded from publicly available sources. This step reduces
dimensionality and allows models to focus on informative words.</p>
        <p>Tokenization is performed using the indic_tokenize module, designed to handle the complexities of
Indian languages. The Tfidf Vectorizer then converts each sentence into meaningful tokens or words and
transforms them into numerical features. By giving words that appear frequently within a document
but rarely across the entire corpus higher weights, Tfidf Vectorizer captures the relative importance
of each word. To maintain computational feasibility, the vocabulary is limited to the top 5,000 most
frequent terms.</p>
        <p>Four machine learning models are employed for sarcasm detection in Tamil text: K-Nearest Neighbors
(KNN), Support Vector Machines (SVM), Decision Trees (DT), and Logistic Regression (LR). These
models are selected for their ability to handle high-dimensional feature spaces and their eficiency in
text classification tasks.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Feature engineering and data splitting</title>
        <p>Initially, a bigger and more varied dataset was produced by combining the initial training and validation
datasets. In order to improve the model’s ability to generalize, more data was made accessible for
training. Following the datasets’ merger, an 80-20 split was implemented, whereby 20% of the merged
data was set aside for testing and 80% was used for training. This method made it possible to train the
model on a greater range of instances while still ofering a distinct test set for assessing how well it
performed on data that wasn’t used for training.</p>
        <p>For feature extraction, Tfidf Vectorizer was utilized to convert text into a numerical matrix, evaluating
the significance of words in relation to the entire dataset. Word embeddings were also employed to
translate words into continuous vector representations, capturing semantic relationships based on
context. To preserve local context and better understand word associations, N-grams, particularly
trigrams, were used. The selection of trigrams was aimed at incorporating patterns involving sequences
of three words, which can provide a deeper understanding of context, especially for capturing nuanced
expressions like sarcasm. The extracted features from these techniques were then used to train and test
the machine learning models.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Model Selection</title>
        <sec id="sec-4-3-1">
          <title>4.3.1. Logistic Regression</title>
          <p>
            A popular classification technique for sarcasm detection, even in contexts with mixed language codes,
is logistic regression. The model uses a probability score between [
            <xref ref-type="bibr" rid="ref1">0,1</xref>
            ] to determine if a text is sarcastic
or not. Texts that meet a predetermined threshold (e.g., 0.5) are classed as sarcastic. The code-mixed
text’s sentiment scores, word frequencies, and context-based signals are among the features that are
utilized to produce predictions. While LR yields interpretable findings, it is not always able to handle
the delicate, context-dependent nature of sarcasm, particularly in texts with mixed codes. Sarcasm
frequently conveys contradicting or weak feelings that are dificult to define with straightforward linear
bounds. We have obtained an accuracy of 75.49% with F1 score of 0.371 in Tamil dataset and an accuracy
of 83.61% in the Malayalam dataset
          </p>
        </sec>
        <sec id="sec-4-3-2">
          <title>4.3.2. K-Nearest Neignbour(KNN)</title>
          <p>By categorizing texts based on their similarity to pre-labeled sarcastic or non-sarcastic samples, the
k-Nearest Neighbors method (KNN) can detect sarcasm without needing complex frameworks. When
applied to code-mixed languages, KNN analyzes features from both languages, such as word frequency,
sentence structure, and contextual patterns, to find the closest matches. The final label is determined by
the majority class of the nearest neighbors.</p>
          <p>Although simple and easy to implement, KNN struggles to accurately identify sarcasm, especially in
code-mixed texts. Traditional similarity metrics often fail to capture subtle cues like tone, wordplay,
or the context of language switching, which are critical for detecting sarcasm. As a result, KNN’s
efectiveness is limited in these scenarios. On the Tamil dataset, we achieved an accuracy of 73.51%
with an F1 score of 0.238. For the Malayalam dataset, the accuracy reached 76.39%. This shows KNN’s
limitations in identifying sarcasm in complex language settings.</p>
        </sec>
        <sec id="sec-4-3-3">
          <title>4.3.3. Support Vector Machines(SVM)</title>
          <p>When it comes to code-mixed languages like Tamil-English, Support Vector Machines (SVM) are
especially helpful in diferentiating between messages that are sarcastic and those that are not. By
converting the text into numerical features, such TF-IDF scores, and extracting linguistic traits unique
to the code-mixed language, SVM finds the ideal hyperplane that divides the two groups. SVM is a good
option since sarcasm in these kinds of writings frequently depends on how both languages interact
with one another and their contextual meanings. It can handle sparsity and high-dimensional feature
spaces well, which makes it appropriate for sarcasm detection in code-mixed, complicated, context-rich
datasets. We have obtained an accuracy of 66.424% with F1 score of 0.321% in Tamil dataset and an
accuracy of 71.019% in the Malayalam dataset</p>
        </sec>
        <sec id="sec-4-3-4">
          <title>4.3.4. Decision Tree</title>
          <p>Decision Trees (DT) in sarcasm detection function by recursively segmenting the data according to
the characteristics that most efectively distinguish sarcastic from non-sarcastic occurrences. These
characteristics for texts with mixed codes could be word choice, punctuation, mood changes, and
grammatical patterns in both languages. The feature providing the most information gain is chosen at
each node in the tree, resulting in an organized sequence of choices that establish categorization.</p>
          <p>Although DTs are easily interpreted and shown, a basic tree structure may not be able to properly
capture the complex and context-dependent nature of sarcasm in code-mixed language. In these
situations, feature engineering and the incorporation of DTs into ensemble models can enhance their
eficacy in sarcasm detection. We have obtained an accuracy of 69.359% with F1 score of 0.403 in Tamil
dataset and an accuracy of 77.17% in the Malayalam dataset.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>The methodology involved comparing the actual results published on the website with the outcomes
produced by each model. This approach facilitated a comprehensive assessment of the models’
performance relative to established benchmarks.Identification of sarcasm is essential, with Logistic Regression
(LR) achieving the highest accuracies of 75.50% for Tamil and 83.62% for Malayalam. Despite its
straightforward nature, LR efectively handled code-mixed data, delivering consistent performance
across both languages. In comparison, K-Nearest Neighbors (KNN) struggled with context
identification in code-switching scenarios, resulting in lower accuracies of 73.51% for Tamil and 76.39% for
Malayalam. Decision Trees (DT) outperformed KNN in capturing context-based signals, while Support
Vector Machines (SVM) demonstrated satisfactory performance with high-dimensional data, although
both models faced dificulties in grasping the nuances of sarcasm. Ultimately, the incorporation of
transliteration-based techniques and tailored machine learning models contributed to improved system
accuracy, highlighting the potential for enhanced sarcasm detection across multilingual datasets.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used ChatGPT in order to: drafting content, grammar
and spelling check, etc. After using this tool/service, the author(s) reviewed and edited the content as
needed and take(s) full responsibility for the publication’s content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. N</given-names>
            , B. B,
            <surname>N. K</surname>
          </string-name>
          , T. Durairaj,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Kumaresan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Rajkumar, Overview of sarcasm identification of dravidian languages in dravidiancodemix@fire2024, in: Forum of Information Retrieval and Evaluation FIRE -</article-title>
          <year>2024</year>
          , DAIICT , Gandhinagar,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. R.</given-names>
            <surname>Sangwan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. K.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Wadhwa, Hybrid deep learning model for sarcasm detection in indian indigenous language using word-emoji embeddings</article-title>
          ,
          <source>ACM Trans. Asian LowResour. Lang. Inf. Process</source>
          .
          <volume>22</volume>
          (
          <year>2023</year>
          ). URL: https://doi.org/10.1145/3519299. doi:
          <volume>10</volume>
          .1145/3519299.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <surname>D. Das</surname>
          </string-name>
          ,
          <article-title>A multimodal approach to sarcasm detection on social media (</article-title>
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>S. M.</given-names>
            <surname>Sarsam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Al-Samarraie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. I.</given-names>
            <surname>Alzahrani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Wright</surname>
          </string-name>
          ,
          <article-title>Sarcasm detection using machine learning algorithms in twitter: A systematic review</article-title>
          ,
          <source>International Journal of Market Research</source>
          <volume>62</volume>
          (
          <year>2020</year>
          )
          <fpage>578</fpage>
          -
          <lpage>598</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>D.</given-names>
            <surname>Vinoth</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Prabhavathy</surname>
          </string-name>
          ,
          <article-title>An intelligent machine learning-based sarcasm detection and classification model on social networks</article-title>
          ,
          <source>The Journal of Supercomputing</source>
          <volume>78</volume>
          (
          <year>2022</year>
          )
          <fpage>10575</fpage>
          -
          <lpage>10594</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>H.</given-names>
            <surname>Thapliyal</surname>
          </string-name>
          ,
          <article-title>Sarcasm Detection System for Hinglish Language (SDSHL)</article-title>
          ,
          <source>Ph.D. thesis, IIIT Hyderabad</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>R.</given-names>
            <surname>Viswanadha</surname>
          </string-name>
          ,
          <article-title>Transliteration of tamil and other indic scripts</article-title>
          ,
          <source>Tamil Internet</source>
          <year>2002</year>
          (
          <year>2002</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>T.</given-names>
            <surname>Ptáek</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Habernal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Hong</surname>
          </string-name>
          ,
          <article-title>Sarcasm detection on czech and english twitter</article-title>
          ,
          <source>in: COLING</source>
          <year>2014</year>
          ,
          <source>the 25th International Conference on Computational Linguistics</source>
          ,
          <year>2014</year>
          , pp.
          <fpage>213</fpage>
          -
          <lpage>223</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>S. K.</given-names>
            <surname>Bharti</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. S.</given-names>
            <surname>Babu</surname>
          </string-name>
          ,
          <article-title>Sarcasm as a contradiction between a tweet and its temporal facts: a pattern-based approach</article-title>
          ,
          <source>International Journal on Natural Language Computing (IJNLC)</source>
          Vol
          <volume>7</volume>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>S. G.</given-names>
            <surname>Shamay-Tsoory</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Tomer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Aharon-Peretz</surname>
          </string-name>
          ,
          <article-title>The neuroanatomical basis of understanding sarcasm and its relationship to social cognition</article-title>
          .,
          <source>Neuropsychology</source>
          <volume>19</volume>
          (
          <year>2005</year>
          )
          <fpage>288</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Dikshit</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. H. C.</given-names>
            <surname>Albuquerque</surname>
          </string-name>
          ,
          <article-title>Explainable artificial intelligence for sarcasm detection in dialogues</article-title>
          ,
          <source>Wireless Communications and Mobile Computing</source>
          <year>2021</year>
          (
          <year>2021</year>
          )
          <fpage>2939334</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>N.</given-names>
            <surname>Sripriya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Durairaj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Nandhini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bharathi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Rajkumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Kumaresan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. C.</given-names>
            <surname>Navaneethakrishnan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <article-title>Findings of shared task on sarcasm identification in code-mixed dravidian languages (</article-title>
          <year>2023</year>
          )
          <fpage>22</fpage>
          -
          <lpage>24</lpage>
          . doi:
          <volume>10</volume>
          .1145/3632754.3633077.
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>O.</given-names>
            <surname>Vitman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kostiuk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Sidorov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gelbukh</surname>
          </string-name>
          ,
          <article-title>Sarcasm detection framework using context, emotion and sentiment features</article-title>
          ,
          <source>Expert systems with applications 234</source>
          (
          <year>2023</year>
          )
          <fpage>121068</fpage>
          -
          <lpage>121068</lpage>
          . doi:
          <volume>10</volume>
          .1016/j.eswa.
          <year>2023</year>
          .
          <volume>121068</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>S.</given-names>
            <surname>Bhosale</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Chaudhuri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. L. R.</given-names>
            <surname>Williams</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Tiwari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dutta</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bhattacharyya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kanojia</surname>
          </string-name>
          ,
          <volume>16</volume>
          .
          <article-title>sarcasm in sight and sound: Benchmarking and expansion to improve multimodal sarcasm detection, arXiv</article-title>
          .org (
          <year>2023</year>
          ). doi:
          <volume>10</volume>
          .48550/arxiv.2310.01430.
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>B. P.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. K.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <article-title>A survey of sarcasm detection techniques in natural language processing</article-title>
          ,
          <year>2023</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>6</lpage>
          . doi:
          <volume>10</volume>
          .1109/ISCON57294.
          <year>2023</year>
          .
          <volume>10112176</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Shao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z. W.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <volume>24</volume>
          . multimodal sarcasm target identification in tweets,
          <year>2022</year>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2022</year>
          .
          <article-title>acl-long</article-title>
          .
          <volume>562</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Sripriya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bharathi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Nandhini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. Chinnaudayar</given-names>
            <surname>Navaneethakrishnan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Durairaj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Kumaresan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Rajkumar, Overview of the shared task on sarcasm identification of Dravidian languages (Malayalam and Tamil) in DravidianCodeMix, in: Forum of Information Retrieval and Evaluation FIRE -</article-title>
          <year>2023</year>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Sripriya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bharathi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Nandhini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. C.</given-names>
            <surname>Navaneethakrishnan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Durairaj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Kumaresan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Rajkumar, Overview of the shared task on sarcasm identification of dravidian languages (malayalam and tamil) in dravidiancodemix, in: Forum of Information Retrieval and Evaluation FIRE-</article-title>
          <year>2023</year>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>N.</given-names>
            <surname>Sripriya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Durairaj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Nandhini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bharathi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Rajkumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Kumaresan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Subalalitha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <article-title>Findings of shared task on sarcasm identification in code-mixed dravidian languages</article-title>
          ,
          <source>FIRE</source>
          <year>2023</year>
          16 (
          <year>2023</year>
          )
          <fpage>22</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>