<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>B. Chen); clzhong@fosu.edu.cn (C. Zhong); kongleilei@fosu.edu.cn (L. Kong)</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>CLEF 2025 JOKER Track: Enhancing Humor-Aware Information Retrieval with Relevance-Aware Classification</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Bo Chen</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>ChangLe Zhong</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>LeiLei Kong</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Foshan University</institution>
          ,
          <addr-line>Foshan, Guangdong</addr-line>
          ,
          <country country="CN">China</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2025</year>
      </pub-date>
      <volume>000</volume>
      <fpage>9</fpage>
      <lpage>0009</lpage>
      <abstract>
        <p>This study investigates the humor-aware information retrieval, within the JOKER Lab at CLEF 2025. This task involves retrieving short humorous texts from a document collection and subsequently identifying texts containing puns. A Enhancing Humor-Aware Information Retrieval with Relevance-Aware Classification approach is introduced, utilizing the retrieval model and the classification model to fulfill task objectives with each model adopting its unique methodology and cooperating to enhance overall performance. This collaborative strategy not only achieves the primary task objectives but also diminishes the task's complexity, rendering it more feasible to implement. Experimental results demonstrate that our method achieved second place ranked by MAP (0.16) on the English dataset, with P@5 at 0.36 and NDCG@5 at 0.41. On the Portuguese dataset, it attained first place in both P@5 (0.44) and NDCG@5 (0.83), while achieving second place in MAP (0.40).. These findings indicate the potential efectiveness of the proposed method. Nevertheless, significant scope for improvement remains, warranting further exploration and research to enhance the overall methodology.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;pun detection</kwd>
        <kwd>humor-aware information retrieval</kwd>
        <kwd>Relevance-Aware Classification</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>The paper is structured as follows: Section 2 reviews recent advancements in pun detection
methodologies. Section 3 details our proposed approach. Section 4 presents and discusses the experimental
results. Concluding remarks and future directions are provided in Section 5.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>Pun detection has been a topic of computational linguistics research for many years[10]. Recent
advancements have primarily leveraged deep learning techniques. A dominant approach frames
pun detection and localization as a sequence labeling task. Notably, BiLSTM-CRF models have been
efectively applied in this context. These models typically employ specialized annotation schemes, such
as NP (Normal Phrase) and BPA (Bi-directional Pun Annotation), where the BPA scheme is particularly
designed to capture structural constraints by ensuring at most one word per context is labeled as a
pun[6, 7].</p>
      <p>Within this BiLSTM-CRF framework, a bidirectional long short-term memory network (BiLSTM)
learns contextual information from the input sequence, while a conditional random field (CRF) layer
handles sequence label prediction and models dependencies between labels. To enhance performance,
these models often incorporate diverse input features. These include pre-trained word embeddings,
character-level features processed through character-level LSTMs and highway networks, and positional
features indicating word locations[6, 7]. By treating both detection (identifying the presence of a pun)
and localization (identifying the pun word) as a unified sequence labeling problem, this approach learns
to perform both tasks concurrently during training, outputting label probabilities for each word and
decoding the final sequence via the CRF layer[6, 7].</p>
      <p>
        More recently, the advent of large language models (LLMs) has opened new avenues for pun detection,
shifting focus towards leveraging their inherent semantic understanding and generation capabilities.
Initial explorations of LLMs for this task, such as using ChatGPT and the SimpleT5 model, were
conducted during CLEF 2023, demonstrating promising results in understanding and identifying puns in
texts[7, 8]. Building on this potential, subsequent research has adapted powerful sequence-to-sequence
transformers like T5 specifically for pun detection[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
      </p>
      <p>
        The T5-based methodology involves several key steps[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. First, queries are augmented by sourcing
synonyms for query terms using resources like WordNet to improve document matching potential. The
augmented queries and document texts are then transformed into embedding vectors using eficient
tokenizers (e.g., all-MiniLM-L6-v2). Similarity scores between query and document embeddings are
computed, and a threshold (e.g., 0.35) is applied to filter relevant documents. Finally, a T5 model (e.g.,
lfan-T5-base) is fine-tuned on task-specific training data to detect puns within the filtered documents[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ].
This approach highlights the trend of utilizing pre-trained LLMs and fine-tuning them for the specific
nuances of pun identification.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Our Method</title>
      <p>Our task aims to extract all relevant and humorous texts from the given text dataset  based on a
query . To achieve this goal, we propose a Enhancing Humor-Aware Information Retrieval with
Relevance-Aware Classification method, which consists of two main steps. The first step involves
constructing a simple, fast, and eficient retrieval model. The purpose of this model is to address the
relevance requirement of the task, without considering whether the text is humorous. Through this
step, we can obtain the set of humor source texts - related to each query . The second step
involves building a binary classification model to further screen the humor source texts - related to
each query  obtained from the first step and merge all texts, ultimately yielding the set of relevant
humorous texts  for all queries .Overall process is shown in Figure 1:</p>
      <sec id="sec-3-1">
        <title>3.1. Information Retrieval Step</title>
        <p>Text retrieval is the process of finding documents relevant to a user’s query from large collections.
We employs a vector space model for relevant document retrieval. Within this model, documents
and queries are treated as points or vectors in a vector space. By calculating the similarity between a
document vector and a query vector, the relevance of the document to the query can be determined.
Similarity measures, such as cosine similarity, are used for this calculation. This approach efectively
transforms the text retrieval problem into a geometric problem within a vector space, enabling eficient
document retrieval.</p>
        <p>In the construction of this vector model, we utilize a pre-trained model called
"paraphrase-multilingualmpnet-base-v2" as the encoding model. This model encodes each query  and the entire text dataset ,
generating vectors for each query  and text  . We then compute the similarity score  between 
and  using the cosine similarity method (Equation 1). If the similarity score  exceeds the threshold  ,
we deem  and  to be similar, and consequently, add  to the humor source text set -.
 = cos( ) =</p>
        <p>· 
‖‖‖ ‖
(1)</p>
        <p>Parameters:
: Query vector corresponding to the query .
 : Text vector corresponding to the text  .
 ·  : Dot product of vectors  and  .
‖‖: Euclidean norm (magnitude) of vector .
‖ ‖: Euclidean norm (magnitude) of vector  .
: Cosine similarity score, which ranges from -1 to 1. A score of 1 indicates identical direction, -1
indicates opposite direction, and 0 indicates orthogonality.</p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Pun Detection Step</title>
        <p>We formalize the identification of desired relevant humorous texts as a binary classification task: given
a pair of texts (,  ), a classification model determines whether they represent a relevant humorous
text pair.</p>
        <p>In machine learning, classification is a supervised learning technique aimed at assigning input data to
predefined categories or classes. It involves training a model on a labeled dataset, where each instance
is associated with a specific class label. Given that the oficial data lacks directly suitable fine-tuning
training data, we first construct a training set  using data from the retrieval phase.</p>
        <p>In the initial step, we acquire the humorous source text - for each query . Subsequently,
leveraging the "query," "docid," "qrel," and other details from dataset , we match the corresponding 
and  , and label the text pairs according to "qrel". The first-stage retrieval returned numerous irrelevant
documents. Consequently, annotating this data to build the training set yielded a large number of
negative samples (label 0), creating dataset imbalance. To mitigate this, we employed a method where
we randomly select negative samples from a large pool, with a quantity 50% higher than the number of
positive samples, to be used together with the positive samples as training data.</p>
        <p>In the construction of the binary classification model, akin to the first step, we employ the pre-trained
"xlm-roberta-base" model as the foundation. We fine-tune this model using the cross-entropy function
(Equation 2) as the objective loss function. Throughout the training process, we retain the model
iteration that achieves the highest F1 score. This top-performing model is then applied to the final
humorous text recognition task, resulting in the desired relevant humorous texts .</p>
        <p>1 ∑︁ [ log ˆ + (1 − ) log(1 − ˆ)]
ℒ = − 
=1
(2)</p>
        <p>Parameters:
ℒ: Cross-entropy loss.
 : Number of training samples.
: True label of the -th sample.
ˆ: Predicted probability of the -th sample.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments And Evaluation</title>
      <sec id="sec-4-1">
        <title>4.1. Data Description</title>
        <p>This year’s dataset not only includes English data but also adds Portuguese data, and the English dataset
has more content than last year’s.tent. For the English dataset, there are a total of 77,658 corpus texts,
with 5,198 being humorous texts and the rest being non-humorous texts but related to certain queries.
There are 12 queries with labeled information that can be used for model training, and 219 queries
that we ultimately need to use for testing the output results.In the Portuguese dataset, the quantity
is significantly smaller than that of the English dataset. There are only 45,126 text entries, with 1,199
being humorous texts. Additionally, there are 98 queries used for generating output results and 29
queries used to assist in model training. The distribution of data across both languages is shown in
Table 1.</p>
        <p>
          Furthermore, apart from the diferences in quantity, the composition structure and information of
the data are consistent. All data files are individual JSON files, and the key ID information is explained
as follows[
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]:
• docid:a unique document identifier
• text: the text of the instance, which may or may not contain wordplay
• qid:a unique query identifier from the input file
• query:the search query
• qrel:indication the document docid is relevant to the query qid and is a wordplay instance
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Experimental Setting</title>
        <p>In the first step, we selected the multilingual pre-trained model paraphrase-multilingual-mpnet-base-v2
as our encoding model for information retrieval. This model is capable of handling both English and
Portuguese simultaneously, thereby reducing the time costs associated with switching between diferent
models. To maximize the retrieval of relevant texts, we employed a simple static threshold method,
setting the threshold  to 0. In the second step, we continued with this approach, choosing an appropriate
multilingual classification pre-trained model, specifically xlm-roberta-base. All experiments were
conducted on a single NVIDIA A800 GPU (80GB memory). We utilized the HuggingFace transformers
library for model training, with AdamW as the optimizer. The learning rate was set to 2e-5, and the
weight decay was 0.1. Training was carried out for 8 epochs with a batch size of 16, and a linear warmup
strategy was applied with 10% warmup ratio. The maximum input sequence length was set to 128.
• paraphrase-multilingual-mpnet-base-v2:The "Paraphrase-Multilingual-MPNet-Base-V2"
model is a powerful tool for generating paraphrases across multiple languages. It’s built on
the MPNet architecture, which combines masked and permuted language modeling to capture
context efectively. The model is trained on multilingual data, allowing it to handle text in various
languages with high accuracy. It is an improved version (V2) of an earlier model, ofering better
performance and more reliable paraphrasing.[9]
• xlm-roberta-base:The "XLM-RoBERTa-Base" model is a multilingual version of the RoBERTa
architecture, designed to handle text in multiple languages. It is trained on a massive multilingual
corpus, enabling it to understand and generate text across diferent languages efectively. This
model uses byte-level BPE tokenization, which helps in better handling of out-of-vocabulary
words and diverse scripts. As a "base" model, it ofers a good balance between computational
eficiency and performance, making it suitable for a wide range of multilingual NLP tasks.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Evaluation Metrics</title>
        <p>
          To directly evaluate the method proposed in the paper, we adopt the oficial evaluation metrics for
assessment, focusing on extracting a subset of these metrics to showcase its performance. [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ]
• map:mean average precision – i.e., the mean of the average precision scores for each query
• ndcg:normalised discounted cumulative gain, the gain of each document based on its
relevance,discounted logarithmically by its position in the ranking normalised over the ideal ranking
• bpref:binary preference, a sum-based metric showing how many relevant documents are ranked
before irrelevant documents
• MRR:mean reciprocal rank, the average of the multiplicative inverse of the ranks of the first
correct answer of results for a sample of queries
• P5, P10, P15:precision – i.e. the ability of a system to present only relevant items, at diferent
numbers of top ranked results
        </p>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. Experimental Results</title>
        <p>This section primarily presents the evaluation metrics of our method across two languages and compares
it with methods proposed by other teams. Table 2 shows the evaluation results for English, while Table
3 displays the evaluation results for Portuguese.</p>
        <p>In the Table 2, our approach demonstrated both strengths and challenges. Compared to the
"Cryptpix_SBERT" team, it showed a slight advantage in Rprec (Recall Precision), with scores of 0.19 versus
0.15. This indicates a relative strength in retrieving relevant documents within the English context.
However, compared to "UAds_team_3", our method was at a disadvantage across key metrics including
map (mean average precision), P@5, P@10, P@100, and NDCG@5. For instance, map was 0.16 vs.
0.13 and P@5 was 0.21 vs. 0.24. This highlights significant room for improvement in overall retrieval
precision, the accuracy of the top 5, 10, and 100 results, and ranking quality. Furthermore, compared
to "UAds_RM3", while our approach held a marginal edge in P@5 (0.21 vs. 0.20), performance was
generally comparable across other metrics: map (0.16 vs. 0.15), P@10 (0.26 vs. 0.24), P@100 (0.36 vs.
0.40), and NDCG@5 (0.41 vs. 0.43). Nevertheless, "UAds_RM3" performed slightly better in the more
critical areas of accuracy for the top 100 results and ranking quality for the top 5 results.
bpref
pjmathematician_Q14-Q4-R
Rasion_SenTransF+Roberta</p>
        <p>UAds_pt_bm25
Skommarkhos_BM25_E5_MiniLM</p>
        <p>pjmathematician_Q06-gist
Skommarkhos_BM25_E5_MiniLM
results_pt_pt_finetuned</p>
        <p>UAds_pt_rm3
myteam_BERT
duth_xanthi_pt
pjmathematician_Q06-gist32</p>
        <p>UAds_pt_rm3_CE1K
UAds_pt_bm25_CE1K
0.42
0.40
0.08
0.07
0.07
0.07
0.07
0.07
0.06
0.06
0.04
0.04
0.04
0.41
0.41
0.06
0.05
0.04
0.05
0.06
0.05
0.06
0.05
0.02
0.02
0.02
0.44
0.44
0.05
0.06
0.05
0.05
0.06
0.09
0.08
0.07
0.07
0.05
0.04
0.34
0.38
0.06
0.06
0.06
0.06
0.07
0.05
0.06
0.08
0.03
0.03
0.01
0.09
0.08
0.03
0.03
0.03
0.03
0.03
0.03
0.03
0.03
0.02
0.02
0.02
0.51
0.51
0.06
0.07
0.08
0.06
0.07
0.04
0.08
0.07
0.07
0.07
0.04
0.58
0.83
0.11
0.08
0.07
0.07
0.32
0.10
0.07
0.15
0.04
0.04
0.04</p>
        <p>Transitioning to the Portuguese-language environment, our approach performed notably well. It
achieved an Rprec score of 0.41, second only to "pjmathematician_Q14-Q-R"’s 0.42. This underscores
its strong capability for retrieving relevant documents in Portuguese. However, the map score was
0.40, slightly lower than "pjmathematician_Q14-Q-R"’s 0.42, indicating potential for improvement
in overall retrieval precision. Performance on P@5, P@10, and P@100 was largely on par with
"pjmathematician_Q14-Q-R". Yet, it fell short in NDCG@5, suggesting room for enhancing ranking
quality.</p>
        <p>Comparing performance across the two languages reveals that our approach performs better overall
in the Portuguese environment. In Portuguese, it achieved relatively high scores on key metrics like
map, Rprec, and P@5. Its Rprec score was particularly close to the leading team, reflecting better
model adaptation to Portuguese datasets and more efective handling of Portuguese text features and
semantic information. In contrast, performance metrics in the English environment were generally lower.
The comparative disadvantage against teams like "UAds_team_3" and "UAds_RM3" suggests potential
shortcomings in handling English vocabulary, the depth of semantic understanding, or optimization on
English training data. Further refinement is needed to enhance retrieval performance in English.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>This study introduces a Enhancing Humor-Aware Information Retrieval with Relevance-Aware
Classiifcation approach, conceptually inspired by the inherent duality of the task requirements. Crucially,
these distinct requirements are addressed separately and processed independently through dedicated
screening modules, prior to their synergistic integration for generating the final output.</p>
      <p>However, the observed suboptimal performance metrics, specifically low precision and recall,
highlight significant areas necessitating improvement. While leveraging state-of-the-art pre-trained models
followed by appropriate fine-tuning undoubtedly yields substantial performance gains, our
investigation reveals that the method encounters persistent and significant challenges primarily within the
data processing pipeline. Given the paramount importance of data quality and quantity for model
eficacy, future research eforts should prioritize exploring efective strategies for data enhancement
under resource-constrained conditions. This includes investigating techniques such as targeted data
augmentation, intelligent sampling, synthetic data generation, or advanced data cleaning methodologies
to mitigate data-related bottlenecks and unlock further performance potential.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This work is supported by the Quality Engineering Projects for Teaching Quality and Teaching Reform
in Undergraduate Colleges and Universities of Guangdong Province (No.xxx).</p>
    </sec>
    <sec id="sec-7">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the author(s) used DeepSeek, Kimi in order to: Grammar and
spelling check, Paraphrase and reword. After using this tool/service, the author(s) reviewed and edited
the content as needed and take(s) full responsibility for the publication’s content.
[6] L. Ren, B. Xu, H. Lin, L. Yang. Abml: attention-based multi-task learning for jointly humor
recognition and pun detection. Soft Computing 25 (2021) 14109–14118.
[7] L. Ermakova, T. Miller, A.-G. Bosser, V. M. Palma Preciado, G. Sidorov, A. Jatowt. Overview
of JOKER – CLEF-2023 track on automatic wordplay analysis. In: A. Arampatzis, E. Kanoulas,
T. Tsikrika, S. Vrochidis, A. Giachanou, D. Li, M. Aliannejadi, M. Vlachos, G. Faggioli, N. Ferro
(Eds.), Experimental IR Meets Multilinguality, Multimodality, and Interaction. Springer Nature
Switzerland, Cham, 2023, pp. 397–415.
[8] Q. Dubreuil. UBO Team @ CLEF JOKER 2023 Track For Task 1, 2 and 3 - Applying AI Models In</p>
      <p>Regards To Pun Translation. 2023. URL: https://ceur-ws.org/Vol-3497/paper-155.pdf.
[9] Wennie Yang, Jieyu Zhao, Guangxuan Wang, Peizhen Lou, Nianzu Zhang, Sinno Jialin Pan.</p>
      <p>Measuring and Reducing Gendered Correlations in Pre-trained Models. In: Proceedings of the
2023 Conference on Fairness, Accountability, and Transparency (FAT* ’23). New York, NY, USA:
Association for Computing Machinery, 2023, pp. 123–134.
[10] Liana Ermakova et al. CLEF 2024 Joker Task 1: Exploring Pun Detection Using The T5 Transformer
Model. In: Working Notes of the Conference and Labs of the Evaluation Forum (CLEF 2024). CEUR
Workshop Proceedings, CEUR-WS.org, 2024.
[11] Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Mandar Joshi, Danqi Chen, Omer Levy, Mike
Lewis, Luke Zettlemoyer, Veselin Stoyanov. RoBERTa: A Robustly Optimized BERT Pretraining
Approach. arXiv preprint arXiv:1907.11692 (2019).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Liana</given-names>
            <surname>Ermakova</surname>
          </string-name>
          , Ricardo Campos,
          <string-name>
            <surname>Anne-Gwenn</surname>
            <given-names>Bosser</given-names>
          </string-name>
          , Tristan
          <string-name>
            <surname>Miller</surname>
          </string-name>
          (
          <year>2025</year>
          ).
          <article-title>Overview of CLEF 2025 Joker Task 1: Humour-aware Information Retrieval</article-title>
          . In: Guglielmo Faggioli, Nicola Ferro, Paolo Rosso, Damiano Spina (Eds).
          <year>2025</year>
          . Working Notes of the Conference and
          <article-title>Labs of the Evaluation Forum (CLEF 2025)</article-title>
          .
          <source>CEUR Workshop Proceedings, CEUR-WS.org.</source>
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Liana</given-names>
            <surname>Ermakova</surname>
          </string-name>
          , Ricardo Campos,
          <string-name>
            <surname>Anne-Gwenn Bosser</surname>
            , and
            <given-names>Tristan</given-names>
          </string-name>
          <string-name>
            <surname>Miller</surname>
          </string-name>
          .
          <article-title>Overview of the CLEF 2025 JOKER Lab: Humour in Machine</article-title>
          . In Jorge
          <string-name>
            <surname>Carrillo-de-Albornoz</surname>
          </string-name>
          , Julio Gonzalo, Laura Plaza, Alba García Seco de Herrera, Josiane Mothe, Florina Piroi, Paolo Rosso, Damiano Spina, Guglielmo Faggioli, and Nicola Ferro, editors,
          <source>Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Sixteenth International Conference of the CLEF Association (CLEF</source>
          <year>2025</year>
          ),
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>B. M.</given-names>
            <surname>Savage</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H. L.</given-names>
            <surname>Lujan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. R.</given-names>
            <surname>Thipparthi</surname>
          </string-name>
          ,
          <string-name>
            <surname>S. E. DiCarlo.</surname>
          </string-name>
          <article-title>Humour, laughter, learning, and health! a brief review</article-title>
          .
          <source>Advances in Physiology Education</source>
          (
          <year>2017</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Arina</given-names>
            <surname>Gepalova</surname>
          </string-name>
          ,
          <string-name>
            <surname>Adrian-Gabriel Chifu</surname>
            , and
            <given-names>Sébastien</given-names>
          </string-name>
          <string-name>
            <surname>Fournier</surname>
          </string-name>
          .
          <article-title>CLEF 2024 Joker Task 1: Exploring Pun Detection Using The T5 Transformer Model</article-title>
          .
          <article-title>Laboratoire d'Informatique et des Systèmes (LIS) UMR 7020</article-title>
          ,
          <string-name>
            <surname>Aix-Marseille</surname>
            <given-names>Université</given-names>
          </string-name>
          , Université de Toulon, CNRS, LIS, France.
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Lu</surname>
          </string-name>
          .
          <article-title>Joint detection and location of english puns</article-title>
          . CoRR abs/
          <year>1909</year>
          .00175 (
          <year>2019</year>
          ). URL: http://arxiv.org/abs/
          <year>1909</year>
          .00175. arXiv:
          <year>1909</year>
          .00175.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>