<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Vayam Solve Kurmaha @ CLEF 2024: Task 2: Humor Classification according to Genre and Technique using BERT Embeddings and Transformers</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Dhannya S M</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lakshmi Priya S</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Shreedevi Seluka Balaji</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sai Nikitha N.S.R</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Shwetha S</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Surabhi Kamath</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Srinidhi Lakshmi Narayanan</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Sri Sivasubramaniya Nadar College of Engineering</institution>
          ,
          <addr-line>Chennai</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <abstract>
        <p>This study addresses humor classification using advanced NLP techniques. By using the LaBSE model classifier, this approach aims to enhance semantic understanding and accuracy in categorizing all the diferent humor types. Our approach utilizes an advanced variant of BERT, to achieve more nuanced and accurate humor classification in textual-based content. This task is more achievable with BERT since it allows understanding of context from both previous and subsequent words in a sentence, leading to a deeper comprehension of context-dependent nuances inherent in humor. Our model ranked 28th among 54 submissions showing the robustness and accurate nature of our approach. The proposed model demonstrates improved performance in the task of humor classification.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;BERT embeddings</kwd>
        <kwd>tokenizers</kwd>
        <kwd>vectors</kwd>
        <kwd>classification</kwd>
        <kwd>wordplay</kwd>
        <kwd>humor</kwd>
        <kwd>sarcasm</kwd>
        <kwd>humor detection</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>Social media platforms have changed the nature of communication by granting previously unheard-of
freedoms of expression and democratizing the production and dissemination of content. But these
advantages have also brought with them dificulties, especially when it comes to policing content,
especially comedy, which frequently has nuanced undertones and context-dependent implications.</p>
      <p>Because humor is subjective and situational, it presents special challenges for categorization and
moderation. Examples of these challenges include satire, sarcasm, and irony. Because of this dificulty,
sophisticated Natural Language Processing (NLP) methods are required to properly categorize amusing
content and promote a more inclusive and secure online community.</p>
      <p>
        This paper delves into classifying humor within textual content, as outlined in the JOKER lab at
CLEF 2024. Leveraging insights from previous works such as those by Ermakova et al. (2024)[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] and
Palma Preciado et al. (2024)[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], this study aims to contribute to the advancement of humor analysis by
employing state-of-the-art NLP techniques.
      </p>
      <p>
        The objectives of this study are to enhance semantic understanding and classification accuracy across
diverse textual datasets, focusing on six specific categories of humor: Sarcasm (SC), Exaggeration (EX),
Wit and Surprise (WS), Incongruity and Absurdity (AID), Irony (IR), and Self-deprecating humor (SD).
Our approach builds upon existing research and methodologies, incorporating novel techniques to
address the intricacies of humor classification [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>The remainder of this paper is organized as follows: Section 2 talks about works related to ours,
and Section 3 describes the dataset and its features. Section 4 delves into the approach and methods
used, detailing the techniques employed for humor classification. Section 5 discusses the results and
the conclusions drawn from it. Finally, Section 6 concludes the paper and suggests avenues for future
research.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Works</title>
      <p>Humor Classification is considered to be one of the more dificult NLP tasks as humor is quite subjective
and this aspect can make it hard to classify. In recent times, humor classification has gained popularity
due to its impact on downstream tasks such as sentiment analysis and opinion mining.</p>
      <p>
        Recent studies, as done by Hossain et al., 2020 [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] show them testing hypotheses for their classification
task by measuring distances between GloVe vectors but fell short when it came to detecting humor
in non-uniformly labeled data. Another study by Meaney et al., 2021 [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] details how many teams
used diferent approaches to detect both ofense and humor yet participating teams did not meet the
threshold that was expected concerning ground truth humor values, making their humor predictions
weak. Zhang and Liu, 2014 [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] recognized humorous content using Federated Learning to make it
more personalized. This approach is not entirely useful as it is highly personalised. Ren et al., 2021 [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]
proposed ABML, an attention-based multi-task learning model that jointly performs humor recognition
and pun detection, demonstrating the efectiveness of multi-task learning approaches in humor-related
tasks, however, the ABML model’s efectiveness in humor detection and pun recognition may be
constrained by its domain-specific performance on datasets like SemEval2017 Task7 and the 16000
one-liner dataset, thereby limiting generalizability across diverse humor types. In a paper published
in 2020 [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], the authors talk about using machine learning approaches like using word embeddings to
detect self-deprecating humor. This falls short in areas where other types of humor need to be detected.
Mihalcea and Strapparava, 2005 [8] made significant contributions to linguistic feature engineering
for humor prediction also known as humor detection. Their approach concentrated on extracting
linguistic features from text, such as word embeddings, syntactic structures, and sentiment analysis, to
capture the information necessary for humor prediction. However, their approach is limited by certain
surface-level features. Reyes et al., 2012 [? ] explored humor detection in tweets by analyzing stylistic
and semantic features, emphasizing the role of irony and sarcasm in humorous texts and yet it may
face limitations due to potential biases in user-generated tags from social media like Twitter, which
could afect the generalizability of humor and irony classification beyond specific cultural and linguistic
contexts. Work done by Faruqi, Shrivastava [9] uses a Neural Language Model called RNN (Recurrent
Neural Network) with an LSTM (Long Short-term Memory) layer to capture long-term dependencies,
but only to classify as humorous and non-humorous. Chiara Bucaria, 2004 [10] analyzed some forms of
lexical and linguistic ambiguity in English in a specific register, i.e. newspaper headlines. However, it
may be limited by its reliance on a specific corpus of newspaper headlines, potentially constraining
the findings of other forms of written or spoken humor beyond this particular genre. Another paper
published by Fahim, Khan, et al. [11] delves into a comparison of classical machine learning models
such as Logistic Regression, Random Forest and Support Vector Machines concluding that an ensemble
model gives them the best accuracy in detecting whether the sentence is humorous or not. Again, this
is just to classify as humorous and non-humorous whereas our task is to detect diferent types of humor
and label them accordingly.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Dataset</title>
      <p>
        The datasets for both the training and test data were provided by the organizers and is a mixture of
existing corpora on irony and sarcasm, COVID-19 humor, public sources as well as the JOKER 2023
corpora [12, 13, 14, 15, 16]. To start, on performing an analysis of the training dataset [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] that was
provided, given below are the observations:
• The training dataset consisted of 2,454 entries.
      </p>
      <p>• Each entry had an ID, the text to be evaluated, and the Class label
• The testing dataset that was given to us consisted of 6,642 entries.
• The Class labels were:
– SC: Sarcasm
– EX: Exaggeration
– WS: Wit and Surprise
– AID: Incongruity and Absurdity
– IR: Irony
– SD: Self-deprecating humor</p>
    </sec>
    <sec id="sec-4">
      <title>4. Approach</title>
      <sec id="sec-4-1">
        <title>4.1. Pre-processing</title>
        <p>Data preprocessing is a crucial step that is done before the data is passed on to the machine learning
model. The raw data is processed to improve the quality and accuracy of our machine-learning model.
Noise refers to any extraneous information in the data that does not help with the performance of our
model. Data preprocessing removes noise from the dataset. This paper’s approach uses the following
data processing steps:
• Converting to lower: All the characters in the text are converted to lowercase to standardize
the text. This makes our analysis case insensitive as now the words ‘Apple’ and ‘apple’ will be
considered the same by the model.
• Removing punctuations: Punctuations constitute noise in the dataset as they often do not carry
much meaning. Punctuations are removed using a list of punctuations from the string library.
• Removing stop words: Stop words are filler words that do not significantly contribute to the
semantics of the text. Words like ‘a’, ’the’, and ‘is’ are examples of stopwords. They occur very
frequently and therefore removing them significantly reduced the number of words making the
dataset more manageable.
• Stemming and Lemmatization: Stemming reduces a word to its base form and lemmatization
reduces it to its dictionary form called a ’lemma’. This helps with reducing the size of tokens
without changing the context of the text while also improving the vocabulary of the model.</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Efects of Preprocessing on Sentence Embedding</title>
        <p>Using the preprocessed text helped us generate embeddings that were more consistent and clear, as
the standardized and cleaned text (lowercase, no punctuation, no stopwords) helped reduce noise and
redundancy in the embeddings. Preprocessing helps pick out only the words that matter, which improves
the semantic richness of the generated embeddings. As discussed earlier, stemming and lemmatization
make the vocabulary smaller and more focused, thus increasing the eficiency and efectiveness of the
model. Embeddings generated using consistent and clean data help the model perform better with new,
unseen data as the model is trained to generalize better, making the model more attuned to performing
better with humor detection.
4.3. BERT
BERT is a state-of-the-art deep learning model that uses a transformer-based neural network to
understand human-like language. It was developed by researchers at Google AI Language in 2018. BERT stands
for Bidirectional Encoder Representations from Transformers. BERT uses the encoder architecture of a
transformer to learn contextual relations between the tokens in the input sequence [17].</p>
        <p>During training, BERT processes an input sequence of tokens and generates a sequence of vectors
corresponding to each token. The vectors provide a contextualized representation of the words. It then
employs tasks like the Masked Language Model and Sentence Prediction to define its training objectives.
After training, the model augmented by a classification layer [ 18] is fine-tuned on task-specific data.
Finally, the model is usable for our classification task. The approach in this work involves setting the
number of training epochs to 3 and using a batch size of 16. Additionally, an early stopping technique
is employed.</p>
        <p>Early stopping is a regularization technique that is used to prevent overfitting of data. It observes
the evaluation metric and stops the training process if it stops improving or starts deteriorating. The
patience parameter tells the model how many epochs to wait after it starts noticing a dip or stagnation
in the evaluation metric. The delta parameter defines the minimum positive change in the metric to
consider it as an improvement. In this model, the early stopping patience is set to 3 and the delta to 0.01.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.4. Feature Extraction</title>
        <p>Feature Extraction using BERT was done using LaBSE, Language Agnostic BERT Sentence Embeddings,
as they generate high-quality embeddings that capture the subtle contextual clues in humor as well as
wordplay.</p>
        <p>Since the LaBSE model has been pre-trained on large corpora of multilingual text, it has learned rich
multilingual text that performs better even though our dataset is monolingual.</p>
        <p>LaBSE is also pre-trained to be better at generalizing across tasks which made it more eficient in
humor classification as humor classification often involves a vast range of linguistic constructs.</p>
        <p>Most importantly, LaBSE provides contextual embeddings that take into account the context of their
surrounding words and sentences. This was a game changer for humor classification as humor is largely
contextual.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
      <p>The model was trained and evaluated with an 80-20 train-test split. The results of the classification on
the test set are as follows:</p>
      <p>The table given below summarizes the performance metrics from the CLEF 2024 JOKER task results,
where our team ranked 28th.</p>
      <sec id="sec-5-1">
        <title>5.1. Comparison of Results</title>
        <p>
          From a comparison of the results obtained from our trained model with the performance metrics from
the CLEF 2024 JOKER task, several observations can drawn. Our model performed better in certain
categories such as WS (Wit and Surprise) and SD (Self-deprecating Humor) however, it fell short in
others like IR (Irony) and EX (Exaggeration). The macro and weighted average precision, recall, and
F1-scores are also marginally lower in our trained model compared to the CLEF 2024 results. However,
the overall accuracy of our trained model was 57.59%, which is equivalent to the 60% accuracy achieved
in the CLEF 2024 JOKER [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ] task.
        </p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Detailed Analysis of Test Results</title>
        <p>Our fine-tuned model using BERT embeddings and transformers with a train batch size of 16, achieved
an overall accuracy of 57.59% on the test set. The model’s performance reflects its ability to classify the
types of humor based on genre and technique. The classification report in Table 1 provides a breakdown
of precision, recall, and F1-score for each humor category.</p>
        <p>The model performed well in identifying Wit and Surprise (WS) and Self-deprecating (SD) humor,
achieving F1-scores of 0.78 and 0.57, respectively. Sarcasm (SC) humor’s F1-score also stood at 0.57. These
results indicate that the model successfully understands and captures linguistic nuances associated with
these humor types. However, the model struggled with categories such as Irony (IR) and Exaggeration
(EX), where F1-scores were notably lower at 0.19 and 0.32, respectively.</p>
      </sec>
      <sec id="sec-5-3">
        <title>5.3. Performance Comparison and Ranking</title>
        <p>Comparing our results with the performance metrics from the CLEF 2024 JOKER [12] task (Table 2), the
model ranked 28th out of 54 submissions. The CLEF 2024 results show higher average precision, recall,
and F1-scores across most categories compared to the analysis performed on the run of the train-test
split performed on the training dataset, as results were calculated on 722 entries of the 6,642 in the
test dataset. Thereby, using this fine-tuned BERT model, predictions were made on the actual dataset,
achieving an overall accuracy of 0.60.</p>
        <p>As mentioned in the overview paper [12], our model did take a hit when it came to making predictions
in the WS (Wit and Surprise) category and the EX (Exaggeration) category. Wit and Surprise is a category
that is a combination of two types of humor, making it dificult to predict [ 12]. Our model performed
very well in the AID (Absurdity and Incongruity) model with an F1 score of 0.78.</p>
        <p>Following the CLEF 2024 JOKER [12] task results, our model’s ranking among submissions has been
incorporated into the abstract of our paper. This ranking reflects our model’s competitive standing
within the field of humor classification using advanced NLP techniques.</p>
        <p>Although there are areas needing improvement, the insights from our analysis pave the way for
future research to refine model performance across diferent humor categories.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Conclusion</title>
      <p>Our approach provides a robust model that works well on humor classification. The results demonstrate
that our approach utilizing BERT embeddings and transformers, specifically the LaBSE model, achieves
competitive performance in humor classification. The high precision and recall values indicate that
the model efectively distinguishes between diferent humor types, including subtle nuances like irony
and wordplay. Future work could include extending this model to train and predict entries in other
languages as this approach uses LaBSE which is language agnostic is pre-trained on multiple languages
and could span cultural contexts.</p>
      <p>Our approach of using BERT for classification along with LaBSE for feature extraction shows that
pre-trained models work well for humor classification tasks such as this.
[8] R. Mihalcea, C. Strapparava, Making computers laugh: Investigations in automatic humor
recognition, in: Proceedings of Human Language Technology Conference and Conference on Empirical
Methods in Natural Language Processing, 2005, pp. 531–538.
[9] F. Faruqi, M. Shrivastava, “is this a joke?”: A large humor classification dataset, in: Proceedings of
the 15th International Conference on Natural Language Processing, 2018, pp. 104–109.
[10] C. Bucaria, Lexical and syntactic ambiguity as a source of humor: The case of newspaper headlines,
Humor-international Journal of Humor Research - HUMOR 17 (2004) 279–309. doi:10.1515/
humr.2004.013.
[11] N. I. Fahim, R. Khan, S. Rahman, N. Akter, M. N. Huda, Humor detection using machine learning
approach, in: 2024 6th International Conference on Electrical Engineering and Information
Communication Technology (ICEEICT), 2024, pp. 1217–1222. doi:10.1109/ICEEICT62016.2024.
10534417.
[12] L. Ermakova, T. Miller, A.-G. Bosser, V. M. P. Preciado, G. Sidorov, A. Jatowt, Overview of
CLEF 2024 JOKER track on automatic humor analysis, in: L. Goeuriot, P. Mulhem, G. Quénot,
D. Schwab, L. Soulier, G. M. D. Nunzio, P. Galuščáková, A. G. S. de Herrera, G. Faggioli, N. Ferro
(Eds.), Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of
the Fifteenth International Conference of the CLEF Association (CLEF 2024), Lecture Notes in
Computer Science, Springer-Verlag, 2024.
[13] L. Ermakova, T. Miller, A. G. Bosser, V. M. P. Preciado, G. Sidorov, A. Jatowt, Overview of
JOKER – CLEF-2023 track on automatic wordplay analysis, in: A. Arampatzis, E. Kanoulas,
T. Tsikrika, S. Vrochidis, A. Giachanou, D. Li, M. Aliannejadi, M. Vlachos, G. Faggioli, N. Ferro
(Eds.), Experimental IR Meets Multilinguality, Multimodality, and Interaction, volume 14163
of Lecture Notes in Computer Science, Springer Nature Switzerland, Cham, 2023, pp. 397–415.
doi:10.1007/978-3-031-42448-9\_26.
[14] S. Frenda, A. Pedrani, V. Basile, S. M. Lo, A. T. Cignarella, R. Panizzon, C. Marco, B. Scarlini, V. Patti,
C. Bosco, D. Bernardi, EPIC: Multi-perspective annotation of a corpus of irony, in: Proceedings of
the 61st Annual Meeting of the Association for Computational Linguistics, volume 1, Association
for Computational Linguistics, 2023, pp. 13844–13857. doi:10.18653/v1/2023.acl-long.774.
[15] F. Galatolo, G. Martino, M. Cimino, C. Tommasi, et al., Dense information retrieval on a latin
digital library via labse and latinbert embeddings (2023).
[16] N. R. Bogireddy, S. Suresh, S. Rai, I’m out of breath from laughing! i think? a dataset of
covid19 humor and its toxic variants, in: Companion Proceedings of the ACM Web Conference
2023, Association for Computing Machinery, New York, NY, 2023, pp. 1004–1013. doi:10.1145/
3543873.3587591.
[17] I. Annamoradnejad, G. Zoghi, Colbert: Using bert sentence embedding in parallel neural networks
for computational humor, Expert Systems with Applications 249 (2024) 123685.
[18] R. Gupta, Bidirectional Encoders to State-of-the-Art: A Review of BERT and Its
Transformative Impact on Natural Language Processing, Informatika. Ekonomika. Upravlenie-Informatics.
Economics. Management 3 (2024) 0311–0320.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>L.</given-names>
            <surname>Ermakova</surname>
          </string-name>
          , A.-G. Bosser,
          <string-name>
            <given-names>T.</given-names>
            <surname>Miller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Thomas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V. M. P.</given-names>
            <surname>Preciado</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Sidorov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Jatowt</surname>
          </string-name>
          ,
          <article-title>CLEF 2024 JOKER Lab: Automatic Humour Analysis</article-title>
          , in: N.
          <string-name>
            <surname>Goharian</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Tonellotto</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          <string-name>
            <surname>He</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Lipani</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>McDonald</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Macdonald</surname>
          </string-name>
          , I. Ounis (Eds.),
          <source>Advances in Information Retrieval</source>
          , Springer Nature Switzerland, Cham,
          <year>2024</year>
          , pp.
          <fpage>36</fpage>
          -
          <lpage>43</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>V. M. P.</given-names>
            <surname>Preciado</surname>
          </string-name>
          , et al.,
          <article-title>Overview of the CLEF 2024 JOKER task 2: Humour classification according to genre and technique</article-title>
          , in: G.
          <string-name>
            <surname>Faggioli</surname>
          </string-name>
          , et al. (Eds.),
          <source>Working Notes of the Conference and Labs of the Evaluation Forum (CLEF</source>
          <year>2024</year>
          ), CEUR Workshop Proceedings, CEUR-WS.org,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>N.</given-names>
            <surname>Hossain</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Krumm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gamon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. A.</given-names>
            <surname>Kautz</surname>
          </string-name>
          , SemEval
          <article-title>-2020 Task 7: Assessing Humor in Edited News Headlines</article-title>
          , CoRR abs/
          <year>2008</year>
          .00304 (
          <year>2020</year>
          ). URL: https://arxiv.org/abs/
          <year>2008</year>
          .00304. arXiv:
          <year>2008</year>
          .00304.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J. A.</given-names>
            <surname>Meaney</surname>
          </string-name>
          , S. Wilson,
          <string-name>
            <given-names>L.</given-names>
            <surname>Chiruzzo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Lopez</surname>
          </string-name>
          , W. Magdy,
          <article-title>SemEval 2021 task 7: HaHackathon, detecting and rating humor and ofense</article-title>
          , in: A.
          <string-name>
            <surname>Palmer</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Schneider</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Schluter</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Emerson</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Herbelot</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          Zhu (Eds.),
          <source>Proceedings of the 15th International Workshop on Semantic Evaluation (SemEval-2021)</source>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Online,
          <year>2021</year>
          , pp.
          <fpage>105</fpage>
          -
          <lpage>119</lpage>
          . URL: https: //aclanthology.org/
          <year>2021</year>
          .semeval-
          <volume>1</volume>
          .9. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2021</year>
          .semeval-
          <volume>1</volume>
          .9.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>X.</given-names>
            <surname>Guo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Xing</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Feng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Nie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Miao</surname>
          </string-name>
          ,
          <article-title>Federated learning for personalized humor recognition</article-title>
          ,
          <source>ACM Trans. Intell. Syst. Technol</source>
          .
          <volume>13</volume>
          (
          <year>2022</year>
          ). URL: https://doi.org/10.1145/ 3511710. doi:
          <volume>10</volume>
          .1145/3511710.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>L.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <article-title>Abml: attention-based multi-task learning for jointly humor recognition and pun detection</article-title>
          ,
          <source>Soft Computing</source>
          <volume>25</volume>
          (
          <year>2021</year>
          )
          <fpage>14109</fpage>
          -
          <lpage>14118</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Kamal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Abulaish</surname>
          </string-name>
          ,
          <article-title>Self-deprecating humor detection: A machine learning approach</article-title>
          , in: L.
          <string-name>
            <surname>-M. Nguyen</surname>
            ,
            <given-names>X.-H.</given-names>
          </string-name>
          <string-name>
            <surname>Phan</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Hasida</surname>
          </string-name>
          , S. Tojo (Eds.),
          <source>Computational Linguistics</source>
          , Springer Singapore, Singapore,
          <year>2020</year>
          , pp.
          <fpage>483</fpage>
          -
          <lpage>494</lpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>