<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Malayalam and Tamil</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Shreyas Karthik</string-name>
          <email>shreyas2310140@ssn.edu.in</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Murari Sreekumar</string-name>
          <email>murari2310237@ssn.edu.in</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kushaal Shyam Potta</string-name>
          <email>kushaalshyam2310513@ssn.edu.in</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Durairaj Thenmozhi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Workshop</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Analysis, Text Analytics</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Sarcasm Identification, Traditional Machine Learning Algorithms</institution>
          ,
          <addr-line>Natural Language Processing, Sentiment</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Sri Sivasubramaniya Nadar College Of Engineering</institution>
          ,
          <addr-line>Rajiv Gandhi Salai (OMR), Kalavakkam 603 110, Tamil Nadu</addr-line>
          ,
          <country country="IN">India</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Sarcasm presents a considerable challenge in the field of sentiment analysis due to its inherently context-dependent nature. On social media, where communications are frequently code-mixed, particularly in Dravidian languages, there is an increasing demand for identifying sarcastic content to ensure efective protection of users from sarcasm disguised as hate or harmful speech. The FIRE 2024 shared task aims to detect sarcasm in Tamil and Malayalam comments collected from the social media platform YouTube. Various traditional machine learning approaches are employed to identify whether the comments contain sarcastic content in Tamil and Malayalam languages. Among these, our logistic regression model achieved a MF Score of 0.68 for Tamil and 0.67 for Malayalam, highlighting its strong fit and capability in addressing the complex nuances of sarcasm detection in code-mixed Dravidian languages.The overall rank we obtained for Tamil and Malayalam set is 7 and 8 respectively.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Sarcasm is the use of irony to mock or convey contempt, often by saying the opposite of what one actually
means. Sarcasm can confuse or alienate those who don’t recognize it, leading to misunderstandings
or feelings of frustration. It often serves as a rhetorical device to mask hateful sentiments. It allows
individuals to express negative opinions while maintaining a façade of humor or irony, making it
dificult for both humans and automated systems to identify the underlying hostility. This phenomenon
is particularly evident in social media, where users can employ sarcasm to subtly convey contempt or
aggression without facing immediate repercussions.[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]
      </p>
      <p>
        It is crucial to address these aspects of sarcasm in social networks and hence Natural Language
Processing research is crucial in providing insights into identifying the comments and classifying them
as Sarcastic and Non-Sarcastic. Computational understanding of natural language has been used in
addressing issues such as sentiment analysis[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], human behaviour detection, fake news detection[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ],
question answering and threat detection across diferent forms of media.
      </p>
      <p>Our research paper presents various innovative solutions contributing to the field of sarcasm
identification in significant ways:
• Optimized Approach: The models used in this research like Support Vector Machines (SVM),
Logistic Regression, and Random Forest have their hyperparameters tuned to their finest level so
that it efectively identifies sarcastic tweets.
• This project can be used for real time applications in social media platforms like Twitter, Instagram,</p>
      <p>Facebook, LinkedIn etc in order to maintain a healthy and safe online environment.</p>
      <p>
        The task that we have performed in FIRE 2024 [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ][
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] is Sarcasm Identification of Dravidian
Languages (Malayalam and Tamil). In this task, the systems have to decide whether the particular
comment is Sarcastic or Not Sarcastic.
      </p>
      <p>CEUR</p>
      <p>ceur-ws.org</p>
      <p>
        In this research paper [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], we have discussed the research works that we have done for the
task. The rest of the paper is organised as follows: Section 2 presents a literature survey explaining
the key theories and concepts, research methodologies and the trends and patterns common in the
ifeld of sarcasm identification. Section 3 describes the diferent datasets used and the task performed.
Section 4 talks about the methodology like preprocessing, lemmatization, vectorization and the various
models used for our task. Section 5 talks about our results and performance analysis with other teams
participating in the task. Finally, in Section 6 we talk about the conclusions and the future prospects of
the research work.
      </p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>Various works in the field of Sarcasm Identification were studied and diverse methodologies and
approach for sarcasm identification and classification were employed to solve this issue. Significant
eforts have been made by researchers around the world to develop annotated datasets and apply
deep learning models such as Convolutional Neural Networks (CNNs) and Recurrent Neural Networks
(RNNs). In addition to these, various transformer based models like ROBERT,Distil-Bert have been used
as they have consistently provided excellent accuracy in identifying and classifying texts.</p>
      <p>
        Krishnan et.al.[
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] undertook a similar research to analyse Tamil and Malayalam code mixed Text
and detect sarcasm.The text came from Youtube videos showcasing movie trailers.The text was split
into 2 categroies Tamil mixed and Malayalam mixed.They tried around 6 diferent models such as MLP
classifier,Random Forest classifier and Logistic Regression .For vectorization they used Count Vectorizer
with n-gram range(1,3) and TF-IDF vectorization . They used a MLP classifier architecture, which had
a solitary hidden layer with 128 neurons.A monitoring system was employed to stop training if no
performance improvement was seen after 5 iterations, thereby reducing the danger of overfitting,and a
maximum of 10 iterations were specified to enable optimal model training.
      </p>
      <p>
        Bamman.et.al.[
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] In this paper based on the tweets provided in twitter they had to perform sarcasm
detection.So in this research they tried to tried to exploit logistic regression/maximum entropy, rather
than sarcasm detection, using feature extraction include lexical cues and other corresponding sentiment;
extralinguistic information of an utterance on Twitter such as the author’s background, audience and
communication environment,etc.They used these features in order to do logistic regression.
      </p>
      <p>Chandra.et.al[11] undertook a similar research to analyse Tamil and Malayalam code mixed Text and
detect sarcasm. The dataset provided to them is a valuable resource for our research, encompassing
codemixed comments and posts in Tamil-English and Malayalam-English, sourced from social media. While
comments and posts may consist of multiple sentences, the dataset predominantly featured an average
sentence length of one. Importantly, each comment and post comes with sentiment polarity annotations,
reflecting real-world scenarios and challenges associated with class imbalance.For research methodology
they used bert-based multilingual trained models.mBERT is built on the transformer architecture which
employs self attenuation mechanisms both in encoder and decoder. These models are pre-trained on vast
text corpora, including Wikipedia, and have a well-established track record of delivering exceptional
performance when fine tuned for various downstream tasks. The architectural design begins with the
model taking a special [CLS] token as input, followed by a sequence of words. This input traverses
through the layers, with each layer applying self-attention mechanisms and forwarding the results to
the subsequent encoder. The output from the final layer of the pre-trained mBERT model serves as
the input to a softmax feedforward neural network, a critical component in classifying statements into
two categories: Sarcastic or Non-Sarcastic. This neural network generates a probability distribution
for each word within the sequence across predefined tags. During prediction, the tag with the highest
associated probability is selected as the predicted tag for each word. In the training phase, they carefully
tuned specific hyperparameters to guide the learning process efectively. These included a learning
rate of 0.01, a batch size of 16, and a maximum of 10 training epochs. These hyperparameters were
meticulously optimized to ensure the model’s proficiency in code-mixed language identification and
sarcasm detection.</p>
      <p>Thenmozhi.et.al[12] used INDIC-Bert and Distil-Bert model on a code mixed Tamil and Malayalam
dataset.f sarcasm in Tamil and Malayalam is meticulously prepared, incorporating annotated data.
The dataset undergoes tokenization following the WordPiece scheme specific to IndicBERT, and it is
subsequently partitioned into distinct subsets for training, validation, and testing. The fine-tuning
phase is executed with predefined hyperparameters and employs a cross-entropy loss function. A
comprehensive evaluation is then conducted, utilizing precision, recall, and F1-score metrics to assess
the model’s proficiency in detecting sarcasm within the contexts of Tamil and Malayalam.For Distil-Bert
model. In the context of DistilBERT fine-tuning, a dataset containing examples of sarcasm in Tamil and
Malayalam is collected and tokenized according to DistilBERT’s subword scheme. Subsequently, the
dataset is partitioned into separate sets for training, validation, and testing. Hyperparameters play a
crucial role in guiding the fine-tuning process, with ongoing validation assessments to monitor the
model’s performance. Evaluation metrics, including precision, recall, and F1-score, are employed to
gauge DistilBERT’s efectiveness in identifying sarcasm within the context of Tamil and Malayalam.
.</p>
      <p>However, the detection and analysis of more subtle, implicit forms of sarcasm remain under-explored.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Dataset Description</title>
      <p>This section provides information about the mixed language data in Tamil and Malayalam including the
details about the dataset and how we prepared it.In our research we used traditional machine learning
models such as Support Vector Machines ,Logistic Regression and Multinomial Naive Bayes and so on.
The Tamil dataset has 29570 posts for training ,6336 posts for validating and 6338 posts for testing.The
Malayalam dataset has 13189 posts for training,2827 posts for validation and 2827 posts for testing.The
dataset contains all three types of code-mixed sentences Inter-Sentential switch, Intra-Sentential switch,
and Tag switching. Most comments were written in native script and Roman script with either Tamil /
Malayalam grammar with English lexicon or English grammar with Tamil / Malayalam lexicon. Some
comments were written in Tamil / Malayalam script with English expressions in between.The objective
of this work is to divide the postings into two categories in the datasets for Tamil and Malayalam:
Sarcastic and Non-sarcastic.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Methodology</title>
      <sec id="sec-4-1">
        <title>4.1. Preprocessing</title>
        <p>We trained the traditional machine learning models such as Support Vector Machine (SVM) [13, 14],
Multionmial Naive Bayes and Logistic Regression on the training dataset, evaluated the models on the
dev dataset and submitted our runs by applying the ML models on the test dataset.
Our first step was to clean the data given in order to improve the performance of the machine learning
models:
1. Converting the text to lowercase: This ensures consistency in text data. By doing this the
vocabulary size is reduced and it reduces the computational requirements.
2. Removing punctuation marks:They often point to external resources that are not relevant to the
context of the text being analyzed.
3. Removing http links and emoticons:These do not contribute to the semantic meaning of the text.
4. Removing twitter mentions like @username
5. Removing stop words in english and as well as Tamil to improve the accuracy of the models .
6. Since the training dataset provided had a lot of other text languages such as Arabic ,Telugu,Hindi,etc
we needed to remove these texts in order to improve the accuracy of the model. In order to
overcome this Polyglot library was used which is a library from python in order to perform nlp
tasks.Using this library the texts were classified into their languages and hence accordingly other
language texts were removed .
7. In the test dataset since other language texts were present google translate library was used
in order to translate the texts into Tamil-English,Malayalam-English respectively . We also
meticulously fine-tuned labels for clarity, transforming ”Sarcastic” and ”Non-sarcastic” into
”1” (sarcasm) and ”0” (non-sarcasm), ensuring that our data aligns perfectly with our binary
classification task.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Count Vectorizer using N-grams</title>
        <p>CountVectorizer is a feature extraction technique in natural language processing (NLP) that converts
text data into numerical vectors, which can be used by machine learning algorithms. When you use
CountVectorizer with n-grams, you’re telling it to consider sequences of words (or characters) of a
specific length, called n-grams, instead of just individual words.</p>
        <p>Example :
1) Unigrams (n=1): Each word is treated as a feature. For example, the sentence ”I love cats” would
be split into ”I”, ”love”, and ”cats”.</p>
        <p>2) Bigrams (n=2): Pairs of consecutive words are treated as features. For example, ”I love cats” would
be split into ”I love” and ”love cats”.</p>
        <p>3) Trigrams (n=3): Sequences of three consecutive words are treated as features. For example, ”I love
cats” would be split into ”I love cats”.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Model Evaluation</title>
        <p>
          We have used three models using hard-hard labels such as Sarcastic and Non-Sarcastic. They are:
1. Support Vector Machines: A supervised machine learning algorithm that we used for classification
and regression tasks. It operates by creating a decision boundary that separates n-dimensional
spaces into classes so that a new data point can be assigned to its relevant category.
2. Logistic Regression: It is a regression model mainly used for classification problems. Logistic
regression models the probability that a given input belongs to a particular class. It uses the
logistic function, also known as the sigmoid function, to map any real-valued number into the
range [
          <xref ref-type="bibr" rid="ref1">0, 1</xref>
          ].
3. Mutlinomial Naive Bayes: Multinomial Naive Bayes is a machine learning algorithm that’s often
used for text classification tasks, like spam detection or sentiment analysis. It’s called ”Naive”
because it assumes that the features (like words in a text) are independent of each other, which is
a simplification. The ”Multinomial” part refers to how it models the data: it counts how often
each word appears in the text and uses these counts to predict the category of the text (like
whether an email is spam or not). It’s simple, fast, and works well when the features are word
frequencies or counts.
        </p>
        <p>For SVM, we have tuned the hyper-parameters like regularization parameter (C) and the kernel
parameters, such as the gamma parameter for the radial basis function (RBF) kernel. For Logistic
Regression, we have tuned hyper-parameters like the regularization strength (often denoted as
C). Regularization techniques such as L1 (lasso) and L2 (ridge) are also tuned to improve model
generalization.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results and Performance Analysis</title>
      <sec id="sec-5-1">
        <title>5.1. Performance Analysis</title>
        <p>Scikit-learn, also known as sklearn, is an open-source, machine learning and data modeling library for
Python. It features various classification, regression and clustering algorithms including support vector
machines, random forests, gradient boosting, k-means and DBSCAN, and is designed to interoperate
with the Python libraries, NumPy and SciPy. The sklearn metrics library also provides the classification
report for evaluation of the performance of the model. The performance is measured using the following
metrics:</p>
        <p>1) MF Score: MF score refers to McFadden’s pseudo-R². The value ranges from 0 to 1, where higher
values indicate better performance of the model.</p>
        <p>The comparison of our team The Three Musketeers with the other teams is represented in the form
of the table below.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Experimental Results</title>
      <p>The below given results are the Macro Average Scores for each models used. The highest score recorded
for Tamil language is 0.68 with Logistic Regression model and highest score recorded for Malayalam
language is 0.67 with Logistic Regression model.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Reflections</title>
      <p>Through this paper we learnt about important methods in the filed of natural language processing
and the steps involved in it .We learnt through this task that Logistic regression is a good model for
text classification.Logistic Regression is a simple linear model, which makes it easy to understand and
interpret.In the context of text classification, each word (or feature) contributes to the classification
decision in a linear fashion, which helps in understanding the importance of specific words in the
text.Text datasets, after being transformed into feature vectors (using methods like TF-IDF), are often
sparse, meaning most of the feature values are zero (most words do not appear in most documents).
Logistic Regression performs well with sparse matrices, making it well-suited for text classification
tasks where sparsity is common.</p>
      <p>Class</p>
    </sec>
    <sec id="sec-8">
      <title>8. Conclusion</title>
      <p>Through the scope of this paper, we have explored traditional models for classifying sarcastic and
non-sarcastic comments using the dataset provided by FIRE 2024, focusing on Tamil and Malayalam
languages. Our findings indicate that Logistic Regression model achieved the highest MF score of
0.68 for Tamil and 0.67 for Malayalam. This research contributes to the field of natural language
processing (NLP) by providing valuable insights into addressing content moderation issues on online
platforms. Moreover, the model can be deployed in real-world applications to monitor and mitigate
sarcastic comments on social platforms. In subsequent research, expanding the model to handle
multiclass classification problems, integrating advanced techniques such as attention mechanisms, and
exploring further preprocessing strategies can enhance its efectiveness. We hope these eforts will
contribute significantly to the moderation and detection of sarcastic comments across various social
media platforms.</p>
    </sec>
    <sec id="sec-9">
      <title>Declaration on Generative AI</title>
      <p>The author(s) have not employed any Generative AI tools.
[11] S. Chanda, A. Mishra, S. Pal, Sarcasm detection in tamil and malayalam dravidian code-mixed
text., in: FIRE (Working Notes), 2023, pp. 336–343.
[12] D. Thenmozhi, Sarcasm detection in dravidian languages using transformer models (2023).
[13] S. L. Salzberg, C4. 5: Programs for machine learning by j. ross quinlan. morgan kaufmann publishers,
inc., 1993, 1994.
[14] T. Pranckevičius, V. Marcinkevičius, Comparison of naive bayes, random forest, decision tree,
support vector machines, and logistic regression classifiers for text reviews classification, Baltic
Journal of Modern Computing 5 (2017) 221.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Frenda</surname>
          </string-name>
          , et al.,
          <article-title>The role of sarcasm in hate speech. a multilingual perspective, in: Proceedings of the doctoral symposium of the xxxivinternational conference of the spanish society for natural language processing</article-title>
          (sepln
          <year>2018</year>
          ), Lloret, E.;
          <string-name>
            <surname>Saquete</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ;
          <article-title>Mart ́ınez-</article-title>
          <string-name>
            <surname>Barco</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ;
          <string-name>
            <surname>Moreno</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <year>2018</year>
          , pp.
          <fpage>13</fpage>
          -
          <lpage>17</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>L.</given-names>
            <surname>Khan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Amjad</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ashraf</surname>
          </string-name>
          , H.-
          <string-name>
            <given-names>T.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gelbukh</surname>
          </string-name>
          ,
          <article-title>Urdu sentiment analysis with deep learning methods</article-title>
          ,
          <source>IEEE access 9</source>
          (
          <year>2021</year>
          )
          <fpage>97803</fpage>
          -
          <lpage>97812</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Khanam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Alwasel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Sirafi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Rashid</surname>
          </string-name>
          ,
          <article-title>Fake news detection using machine learning approaches</article-title>
          ,
          <source>in: IOP conference series: materials science and engineering</source>
          , volume
          <volume>1099</volume>
          ,
          <string-name>
            <given-names>IOP</given-names>
            <surname>Publishing</surname>
          </string-name>
          ,
          <year>2021</year>
          , p.
          <fpage>012040</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Sripriya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bharathi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Nandhini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. C.</given-names>
            <surname>Navaneethakrishnan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Durairaj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Kumaresan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Rajkumar, Overview of the shared task on sarcasm identification of dravidian languages (malayalam and tamil) in dravidiancodemix, in: Forum of Information Retrieval and Evaluation FIRE-</article-title>
          <year>2023</year>
          ,
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <article-title>Hope speech detection in youtube comments</article-title>
          ,
          <source>Social Network Analysis and Mining</source>
          <volume>12</volume>
          (
          <year>2022</year>
          )
          <fpage>75</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. N</given-names>
            , B. B,
            <surname>N. K</surname>
          </string-name>
          , T. Durairaj,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Kumaresan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Rajkumar, Overview of sarcasm identification of dravidian languages in dravidiancodemix@fire2024, in: Forum of Information Retrieval and Evaluation FIRE -</article-title>
          <year>2024</year>
          , DAIICT , Gandhinagar,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Hande</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Kumaresan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Priyadharshini</surname>
          </string-name>
          ,
          <article-title>How can we detect homophobia and transphobia? experiments in a multilingual code-mixed setting for social media governance</article-title>
          ,
          <source>International Journal of Information Management Data Insights</source>
          <volume>2</volume>
          (
          <year>2022</year>
          )
          <fpage>100119</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>N.</given-names>
            <surname>Sripriya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Durairaj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Nandhini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bharathi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Rajkumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Kumaresan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Subalalitha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <article-title>Findings of shared task on sarcasm identification in code-mixed dravidian languages</article-title>
          ,
          <source>FIRE</source>
          <year>2023</year>
          16 (
          <year>2023</year>
          )
          <fpage>22</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>D.</given-names>
            <surname>Krishnan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Dharanikota</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bharathi</surname>
          </string-name>
          ,
          <article-title>Cross-linguistic sarcasm detection in tamil and malayalam: A multilingual approach</article-title>
          .,
          <source>in: FIRE (Working Notes)</source>
          ,
          <year>2023</year>
          , pp.
          <fpage>259</fpage>
          -
          <lpage>269</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>D.</given-names>
            <surname>Bamman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Smith</surname>
          </string-name>
          ,
          <article-title>Contextualized sarcasm detection on twitter</article-title>
          ,
          <source>in: proceedings of the international AAAI conference on web and social media</source>
          , volume
          <volume>9</volume>
          ,
          <year>2015</year>
          , pp.
          <fpage>574</fpage>
          -
          <lpage>577</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>