<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Leveraging Machine Learning and BERT for Sarcasm Detection in Text</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Kogilavani Shanmugavadivel</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Priyadharshini C</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Varshini L</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sathyaa S</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of AI, Kongu Engineering College</institution>
          ,
          <addr-line>Perundurai, Erode</addr-line>
        </aff>
      </contrib-group>
      <abstract>
        <p>Sarcasm detection in Tamil-English code-mixed text presents a unique challenge, particularly when traditional machine learning models are employed. This paper explores the application of conventional algorithms such as random forest, logistic regression, and naive Bayes, as well as the transformer-based BERT model. Performance evaluation uses four datasets, focusing on key metrics such as accuracy, precision, recall, and F1-score. BERT demonstrates superior performance, efectively capturing contextual nuances in sarcasm detection, making it a more viable approach for multilingual and code-mixed environments. Future work may expand on these ifndings by utilizing advanced transformer architectures and incorporating additional features like sentiment and emoji-based cues.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Sarcasm detection</kwd>
        <kwd>Machine learning</kwd>
        <kwd>BERT</kwd>
        <kwd>Code-mixed text</kwd>
        <kwd>Tamil-English</kwd>
        <kwd>Transformer models</kwd>
        <kwd>Natural Language Processing</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. INTRODUCTION</title>
    </sec>
    <sec id="sec-2">
      <title>2. LITERATURE SURVEY</title>
      <p>
        BERT-based models have gained significant traction, as demonstrated by Khatri et al. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] who applied
BERT and GloVe embeddings to Twitter data, highlighting the superior performance of pre-trained
transformers in sarcasm detection.
      </p>
      <p>
        Multitask learning has also been explored, with Majumder et al. [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] focusing on combining sentiment
analysis and sarcasm detection, showing how shared learning improves performance in both areas.
      </p>
      <p>
        Similarly, Hiai and Shimada [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] employed Recurrent Neural Networks (RNNs) with relation vectors,
enabling more nuanced detection of sarcastic patterns.
      </p>
      <p>
        Hybrid models, such as those developed by Jain et al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] combined CNN and LSTM for enhanced
sarcasm detection, efectively handling complex text structures.
      </p>
      <p>
        Kumar and Garg [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] compared traditional machine learning models with advanced deep learning
architectures, finding that deep models like BERT outperformed others in accuracy and generalization.
      </p>
      <p>
        Multimodal approaches have further advanced the field. Poria et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] integrated textual, visual,
and audio features to detect sarcasm, achieving better results in multimedia contexts.
      </p>
      <p>
        Ghosh et al. [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ] tackled sarcasm detection in code-mixed languages, addressing the challenges
of detecting sarcasm in multilingual conversations and showing how BERT can be efective in such
contexts.
      </p>
      <p>
        Zhang et al. [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] implemented BERT for sarcasm detection in tweets, emphasizing the role of context
and the efectiveness of transformer models in processing social media language.
      </p>
      <p>
        Eke et al. [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] introduced a context-based feature extraction technique for sarcasm identification
using deep learning models, specifically leveraging BERT for improved accuracy on benchmark datasets.
Their approach highlighted the importance of contextual understanding in sarcasm detection tasks.
      </p>
      <p>
        Pandey and Singh [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] developed a BERT-LSTM model to detect sarcasm in code-mixed social media
posts. The study addressed challenges posed by mixed languages, demonstrating that BERT’s contextual
embeddings combined with LSTM efectively capture sarcasm nuances.
      </p>
      <p>
        Sandor and Babac [
        <xref ref-type="bibr" rid="ref13">13</xref>
        ] used machine learning techniques to detect sarcasm in online comments.
Their work focused on building efective classifiers for sarcasm detection, leveraging multiple feature
extraction methods to handle the complexity of sarcastic language online.
      </p>
      <p>
        Kumar and Sarin [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] introduced WELMSD, a sarcasm detection approach combining word
embeddings and language models. Their study demonstrated the efectiveness of this hybrid model in handling
sarcasm through contextual understanding and word representations.
      </p>
      <p>
        Goel et al. [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ] explored sarcasm detection using deep learning and ensemble learning techniques.
Their approach combined multiple models to improve classification performance, showcasing the
potential of ensemble methods in enhancing sarcasm detection accuracy in multimedia applications.
      </p>
      <p>
        Jeremy et al. [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] investigated sarcasm detection by optimizing various input methods in text data.
Their work highlights the impact of input representations on model performance, emphasizing the
importance of preprocessing and feature engineering in improving sarcasm classification accuracy.
      </p>
      <p>
        Chakravarthi et al. [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] addressed sarcasm detection in Dravidian languages using the
DravidianCodeMix dataset. They explored machine learning and deep learning methods, focusing on overcoming
challenges in code-mixed and low-resource text.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. METHODOLOGY</title>
      <p>This section describes the approach taken to identify sarcasm in code-mixed Tamil-English text by
utilising the transformer-based BERT model in addition to conventional machine learning models. Data
preprocessing, model implementation, and evaluation are some of the crucial steps in the process.</p>
      <sec id="sec-3-1">
        <title>3.1. DATA COLLECTION AND DATA PREPROCESSING</title>
        <p>The dataset for sarcasm detection is composed of multiple files, each serving a specific purpose in the
analysis process. The first dataset, ‘Change Makers Tamil.csv‘, contains 47,960 rows and two columns:
‘Id‘ and ‘Labels‘. The ‘Labels‘ column identifies whether the content is sarcastic or non-sarcastic. This
dataset provides labels necessary for training models on identifying sarcasm.</p>
        <p>The second dataset, ‘sarcasm tam dev.csv‘, includes text data in a multilingual format (Tamil and
English), with two columns: ‘Text‘ and ‘Labels‘. The ‘Text‘ column contains the actual social media
content collected from platforms like Facebook and Twitter, and the ‘Labels‘ column classifies each text
as either sarcastic or non-sarcastic.</p>
        <p>The third dataset, ‘sarcasm tam test without labels.csv‘, contains only the ‘ID‘ and ‘Text‘ columns,
representing the test set where labels are not provided. This dataset will be used for evaluating the
model’s ability to predict sarcasm.</p>
        <p>The fourth dataset, ’sarcasm tam train.csv’, comprises both ’Text’ and ’Labels’ columns , serving as
the training data for the model. This dataset will be utilized to train the model to diferentiate between
sarcastic and non-sarcastic text, ensuring it learns the patterns for sarcasm detection in the Tamil
language.</p>
        <p>In preprocessing, several steps were applied to clean and prepare the text for model training.
Tokenization was performed while retaining both Tamil and English tokens. Text was converted to
lowercase, and transliterated Tamil words were normalized. Special characters, digits, and punctuation
were removed to reduce noise in the data. Additionally, stopwords from both Tamil and English were
ifltered out to focus on meaningful content.</p>
        <p>Sequence padding was applied to ensure uniform input lengths for the models. Contextual embeddings
were generated using BERT to capture nuanced meaning from the text, while more traditional models
utilized GloVe and Word2Vec embeddings. Finally, the dataset was split into training and testing sets
(80-20 split), with cross-validation applied to ensure robust model evaluation. This preprocessing
pipeline ensured the data was clean, consistent, and ready for sarcasm detection models.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Bidirectional Encoder Representation from Transformers</title>
        <p>BERT (Bidirectional Encoder Representations from Transformers) can interpret context and pick up on
little linguistic clues, it is essential for sarcasm detection. In contrast to conventional models, BERT
makes use of a transformer-based architecture with a bidirectional attention mechanism, which allows
it to take into account the words that come before and after one another in a phrase at the same time.
This bi-directional feature is essential for sarcasm recognition, particularly in code-mixed Tamil-English
language where sarcastic statements must be understood in context.</p>
        <p>Since sarcasm frequently relies on opposing literal and intended interpretations, BERT’s pre-trained
embeddings capture deep contextual links between words, which makes it particularly successful in
identifying sarcasm. By honing BERT’s performance on the sarcasm detection task, one can help it
better distinguish between sarcastic and non-sarcastic comments by teaching it task-specific subtleties
in code-mixed text. In comparison to standard models, the BERT-based model outperforms them in
capturing the complexity of multilingual sarcasm, as evidenced by its better accuracy and F1-score.
Performance measurements show that BERT performs significantly better than models such as Random
Forest and Naive Bayes. The BERT classification report is shown in Table 1.</p>
        <p>Beyond its eficiency, BERT’s adaptability makes it scalable and suitable for real-world applications
since it can be optimised for a variety of natural language processing jobs. BERT is a recommended
option for text classification tasks like sarcasm detection because of its strong design and capacity to
handle large-scale datasets.</p>
        <p>Formula : In the BERT architecture, parameters are the quantity of attention heads, hidden units,
and layers. BERT is incredibly successful in sarcasm recognition because the embedding layer uses
token, positional, and segment embeddings to capture word associations and attention layers model
contextual dependencies.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Random Forest</title>
        <p>By aggregating the predictions of several decision trees, Random Forest is an ensemble learning
technique that is important for sarcasm identification. To increase generalisation and decrease overfitting,
each tree in the forest is trained using a diferent random subset of the features and data. As they
combine the results from diferent trees to produce more reliable and precise predictions, Random
Forests do exceptionally well in complicated classification tasks.</p>
        <p>Random Forest uses its capacity to identify patterns in the data to detect sarcasm by taking into
account several word attributes including frequency and word pairings (n-grams). Despite its reliance
on surface-level features such as TF-IDF vectors and bag-of-words, Random Forest is a strong foundation
for sarcasm detection, especially in dificult code-mixed Tamil-English language, since it can handle the
non-linear correlations between these features.</p>
        <p>The Random Forest model struggles to capture the deeper contextual details that are essential for
sarcasm detection, which makes it less accurate than deep learning techniques like BERT. Nevertheless,
it still achieves a respectable level of accuracy. However, it is a useful tool in many text classification
jobs due to its ease of use, interpretability, and capacity to handle big datasets with little adjustment.The
Random Forest model’s classification report is displayed in Table 2.</p>
        <p>Formula : The number of trees in the forest and the depth of each tree determine how many parameters
there are in Random Forest. The model can achieve better generalisation than a single decision tree
since each decision tree is constructed using a portion of the data and characteristics, and the final
result is the majority vote or average of all tree forecasts.</p>
      </sec>
      <sec id="sec-3-4">
        <title>3.4. Logistic Regression</title>
        <p>Logistic Regression is a popular machine learning model,because of its interpretability and efectiveness
in binary classification problems, for sarcasm detection. It is straightforward yet efective. It works by
estimating the likelihood that an input, given its attributes and the goal label, would fall into a particular
class—in this example, sarcastic or non-sarcastic. The output probabilities of logistic regression are
subjected to a sigmoid function before being thresholded to provide binary predictions.</p>
        <p>Word frequencies, n-grams, and TF-IDF vectors are some of the variables that Logistic Regression
utilises to categorise text in order to detect sarcasm in Tamil-English code-mixed text. The deeper
contextual and non-linear linkages that are frequently essential for sarcasm identification can be dificult
for Logistic Regression to capture, even though it is excellent at handling linearly separable data. This
is especially true in multilingual literature where meaning is dependent on nuanced linguistic cues.</p>
        <p>Because of its simplicity and eficiency, Logistic Regression performs fairly well as a baseline model
despite its limits in addressing complicated patterns; it achieves moderate accuracy in many classification
tasks. It is especially helpful for short assessments and in situations where processing power is scarce.The
classification report for logistic regression is displayed in Table 3.</p>
        <p>Formula : The weights assigned to each feature in logistic regression are the parameters, and they
are discovered by gradient descent in order to minimise the log loss function. By reflecting each
feature’s contribution to the classification, these weights enable the model to forecast based on the
linear relationship between the target output and the input data.</p>
      </sec>
      <sec id="sec-3-5">
        <title>3.5. Naive Bayes</title>
        <p>Because of its efectiveness and simplicity, the probabilistic machine learning model Naive Bayes is
frequently employed for sarcasm detection. The model is computationally eficient because it applies
Bayes’ Theorem under the assumption that all characteristics (words) are independent, which is rarely
the case in real-world data. By multiplying the conditional probabilities of each word in the text with
a class label, Naive Bayes determines if a text is sardonic or not. For text classification tasks such as
sarcasm detection, it frequently achieves good results, despite the naive independence assumption.</p>
        <p>Word frequencies and TF-IDF scores are two examples of features that Naive Bayes utilises to classify
text in sarcasm detection for Tamil-English code-mixed text. It performs best when sarcasm is expressed
using particular terms or patterns that can be statistically distinguished from non-sarcastic classes.
However, compared to more sophisticated models like BERT, its independence assumption restricts its
capacity to capture intricate contextual links, making it less successful for subtle sarcasm detection.</p>
        <p>Naive Bayes is a good choice for first phase classification jobs since it is simple to use, computationally
quick, and performs quite well on limited datasets or as a baseline model.The Naive Bayes classification
report is displayed in Table 4.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. RESULTS AND DISCUSSIONS</title>
      <p>The results reveal that the BERT model significantly outperforms traditional machine learning models
Random Forest, Logistic Regression, and Naive Bayes in detecting sarcasm in Tamil English code mixed
text. BERT achieved higher accuracy, precision, recall, and F1-score across four datasets, showcasing its
ability to capture the nuanced context and cultural subtleties of sarcasm. In contrast, traditional models,
which rely on surface level features, struggled to recognize these complexities. The success of BERT can
be attributed to its transformer architecture, which employs bi-directional attention to analyze word
relationships in context. This indicates that advanced models like BERT are more efective for sarcasm
detection in multilingual environments, and future eforts should consider incorporating features like
sentiment analysis and emoji recognition for further improvements.</p>
    </sec>
    <sec id="sec-5">
      <title>5. CONCLUSION</title>
      <p>In conclusion, the findings highlight the efectiveness of BERT for sarcasm detection in Tamil-English
code-mixed text, demonstrating its superiority over traditional machine learning algorithms. The results
emphasize the need for advanced models that can understand contextual and cultural nuances. Future
research should explore additional transformer architectures and incorporate features such as sentiment
analysis to enhance detection accuracy even further. The insights gained can significantly improve
sarcasm detection systems in multilingual contexts.</p>
    </sec>
    <sec id="sec-6">
      <title>6. FUTURE RESEARCH DIRECTIONS</title>
      <p>Future research on sarcasm detection should explore advanced transformer architectures beyond BERT,
such as RoBERTa and GPT, to enhance context understanding. Incorporating multimodal data, including
audio and visual cues, can improve detection accuracy by capturing tone and facial expressions.
Additionally, expanding datasets to include diverse languages and dialects will enhance model generalization.
Integrating sentiment analysis and emoji recognition can refine sarcasm detection further. Finally,
developing real-time systems for social media monitoring and conversational agents could translate
research findings into practical applications, improving understanding of human communication.</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used ChatGPT in order to: drafting content, grammar
and spelling check, etc. After using this tool/service, the author(s) reviewed and edited the content as
needed and take(s) full responsibility for the publication’s content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Sripriya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Bharathi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Nandhini</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. C.</given-names>
            <surname>Navaneethakrishnan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Durairaj</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Kumaresan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Rajkumar, Overview of sarcasm identification of dravidian languages in dravidiancodemix@ fire-2023</article-title>
          , in: FIRE (Working Notes),
          <year>2023</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. R.</given-names>
            <surname>Sangwan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Arora</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Nayyar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Abdel-Basset</surname>
          </string-name>
          , et al.,
          <article-title>Sarcasm detection using soft attention-based bidirectional long short-term memory model with convolution network</article-title>
          ,
          <source>IEEE access 7</source>
          (
          <year>2019</year>
          )
          <fpage>23319</fpage>
          -
          <lpage>23328</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>A.</given-names>
            <surname>Khatri</surname>
          </string-name>
          , et al.,
          <article-title>Sarcasm detection in tweets with bert and glove embeddings</article-title>
          , arXiv preprint arXiv:
          <year>2006</year>
          .
          <volume>11512</volume>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>N.</given-names>
            <surname>Majumder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Poria</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Peng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Chhaya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Cambria</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Gelbukh</surname>
          </string-name>
          ,
          <article-title>Sentiment and sarcasm classification with multitask learning</article-title>
          ,
          <source>IEEE Intelligent Systems</source>
          <volume>34</volume>
          (
          <year>2019</year>
          )
          <fpage>38</fpage>
          -
          <lpage>43</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Hiai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Shimada</surname>
          </string-name>
          ,
          <article-title>Sarcasm detection using rnn with relation vector</article-title>
          ,
          <source>International Journal of Data Warehousing and Mining (IJDWM) 15</source>
          (
          <year>2019</year>
          )
          <fpage>66</fpage>
          -
          <lpage>78</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>D.</given-names>
            <surname>Jain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kumar</surname>
          </string-name>
          , G. Garg,
          <article-title>Sarcasm detection in mash-up language using soft-attention based bi-directional lstm and feature-rich cnn</article-title>
          ,
          <source>Applied Soft Computing</source>
          <volume>91</volume>
          (
          <year>2020</year>
          )
          <fpage>106198</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kumar</surname>
          </string-name>
          , G. Garg,
          <article-title>Empirical study of shallow and deep learning models for sarcasm detection using context in benchmark datasets</article-title>
          ,
          <source>Journal of ambient intelligence and humanized computing 14</source>
          (
          <year>2023</year>
          )
          <fpage>5327</fpage>
          -
          <lpage>5342</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S.</given-names>
            <surname>Poria</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Hazarika</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Majumder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Mihalcea</surname>
          </string-name>
          ,
          <article-title>Beneath the tip of the iceberg: Current challenges and new directions in sentiment analysis research</article-title>
          ,
          <source>IEEE transactions on afective computing 14</source>
          (
          <year>2020</year>
          )
          <fpage>108</fpage>
          -
          <lpage>132</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>S.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ghosh</surname>
          </string-name>
          ,
          <string-name>
            <surname>D. Das</surname>
          </string-name>
          ,
          <article-title>Sentiment identification in code-mixed social media text</article-title>
          ,
          <source>arXiv preprint arXiv:1707.01184</source>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , D. Ma, P. Tiwari,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Masud</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Shorfuzzaman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <article-title>Stance-level sarcasm detection with bert and stance-centered graph attention networks</article-title>
          ,
          <source>ACM Transactions on Internet Technology</source>
          <volume>23</volume>
          (
          <year>2023</year>
          )
          <fpage>1</fpage>
          -
          <lpage>21</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>C. I.</given-names>
            <surname>Eke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Norman</surname>
          </string-name>
          , L. Shuib,
          <article-title>Context-based feature technique for sarcasm identification in benchmark datasets using deep learning and bert model</article-title>
          ,
          <source>IEEE Access 9</source>
          (
          <year>2021</year>
          )
          <fpage>48501</fpage>
          -
          <lpage>48518</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>R.</given-names>
            <surname>Pandey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. P.</given-names>
            <surname>Singh</surname>
          </string-name>
          ,
          <article-title>Bert-lstm model for sarcasm detection in code-mixed social media post</article-title>
          ,
          <source>Journal of Intelligent Information Systems</source>
          <volume>60</volume>
          (
          <year>2023</year>
          )
          <fpage>235</fpage>
          -
          <lpage>254</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>D.</given-names>
            <surname>Šandor</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. B.</given-names>
            <surname>Babac</surname>
          </string-name>
          ,
          <article-title>Sarcasm detection in online comments using machine learning</article-title>
          ,
          <source>Information Discovery and Delivery</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>P.</given-names>
            <surname>Kumar</surname>
          </string-name>
          , G. Sarin,
          <article-title>Welmsd-word embedding and language model based sarcasm detection</article-title>
          ,
          <source>Online Information Review</source>
          <volume>46</volume>
          (
          <year>2022</year>
          )
          <fpage>1242</fpage>
          -
          <lpage>1256</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>P.</given-names>
            <surname>Goel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Jain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Nayyar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Singhal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Srivastava</surname>
          </string-name>
          ,
          <article-title>Sarcasm detection using deep learning and ensemble learning</article-title>
          ,
          <source>Multimedia Tools and Applications</source>
          <volume>81</volume>
          (
          <year>2022</year>
          )
          <fpage>43229</fpage>
          -
          <lpage>43252</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>N.</given-names>
            <surname>Jeremy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. T.</given-names>
            <surname>Tirtawijaya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. S.</given-names>
            <surname>Lumentut</surname>
          </string-name>
          ,
          <article-title>Optimizing sarcasm detection through various input methods in text data</article-title>
          ,
          <source>in: 2024 2nd International Conference on Technology Innovation and Its Applications (ICTIIA)</source>
          , IEEE,
          <year>2024</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>5</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>B. R.</given-names>
            <surname>Chakravarthi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. N</given-names>
            , B. B,
            <surname>N. K</surname>
          </string-name>
          , T. Durairaj,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. K.</given-names>
            <surname>Kumaresan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. K.</given-names>
            <surname>Ponnusamy</surname>
          </string-name>
          ,
          <string-name>
            <surname>C.</surname>
          </string-name>
          <article-title>Rajkumar, Overview of sarcasm identification of dravidian languages in dravidiancodemix@fire2024, in: Forum of Information Retrieval and Evaluation FIRE -</article-title>
          <year>2024</year>
          , DAIICT , Gandhinagar,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>