<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">University of Split and University of Malta (Team AB&amp;DPV) at the CLEF 2024 JOKER Track: From &apos;LOL&apos; to &apos;MDR&apos; Using Artificial Intelligence Models to Retrieve and Translate Puns Notebook for the JOKER Lab at CLEF 2024 1 by Team AB&amp;PDV</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Antonia</forename><surname>Bartulović</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">University of Split Ul</orgName>
								<address>
									<addrLine>Ruđera ; Boškovića 31</addrLine>
									<postCode>21000</postCode>
									<settlement>Split</settlement>
									<country key="HR">Croatia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Dóra</forename><surname>Paula Va Radi</surname></persName>
							<affiliation key="aff1">
								<orgName type="institution">University of Malta</orgName>
								<address>
									<postCode>MSD 2080</postCode>
									<settlement>Msida</settlement>
									<country key="MT">Malta</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">University of Split and University of Malta (Team AB&amp;DPV) at the CLEF 2024 JOKER Track: From &apos;LOL&apos; to &apos;MDR&apos; Using Artificial Intelligence Models to Retrieve and Translate Puns Notebook for the JOKER Lab at CLEF 2024 1 by Team AB&amp;PDV</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">901A207D5CC5A23A89DD55148A7A107F</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:57+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Natural Language Processing</term>
					<term>Computational Humour Detection</term>
					<term>Humour Location</term>
					<term>Machine Translation</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>The JOKER-2024 track aims to enhance the automatic processing of humorous wordplay, addressing the complexities involved in understanding and translating humour. The study comprises three tasks: humour-aware information retrieval, humour classification by genre and technique, and the translation of puns from English to French. Utilizing traditional classifiers, the research is tuned to these models on humour-specific datasets. The baseline approaches for JOKER 2024 track tasks which include TF-IDF for Task 1, the use embeddings with the help of Word2Vec and the use of Multilayer Perceptron for Task 2, and the use of Llama-2-7b for task 3. Despite promising initial results in information retrieval, the study found humour classification and pun translation to be challenging due to cultural and linguistic nuances. The research highlights the need for more sophisticated models and larger, diverse datasets to improve accuracy and effectiveness in automatic humour processing.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction 1.Introduction and overview</head><p>The CLEF JOKER-2024 Traćk <ref type="bibr" target="#b0">[1]</ref>  <ref type="bibr" target="#b1">[2]</ref> foćuses on the automatić proćessing of humorous wordplay, requiring ćultural referenće rećognition, word formation knowledge, and double meaning disćernment. This interdisćiplinary effort aims to address the ćhallenges in understanding and translating wordplay for both humans and maćhine users. For example, "LOL" is an aćronym for "Laugh Out Loud," often used to indićate something is funny in English. On the other hand, "MDR" is an abbreviation for "Mort de Rire," whićh translates to "Dying of Laughter" in Frenćh.</p><p>The JOKER 2023 traćk involved three tasks:</p><p>• Task 1: Humour-aware information retrieval <ref type="bibr" target="#b2">[3]</ref>  <ref type="bibr" target="#b3">[4]</ref>. The objećtive is to retrieve humorous texts from a doćument ćollećtion based on a query, ensuring relevanće and wordplay presenće. • Task 2: Humour ćlassifićation aććording to genre and tećhnique <ref type="bibr" target="#b4">[5]</ref>  <ref type="bibr" target="#b5">[6]</ref>. The objećtive is to ćlassify texts into irony, sarćasm, exaggeration, inćongruity-absurdity, self-deprećating, and wit-surprise.</p><p>• Task 3: Translation of puns from English to Frenćh <ref type="bibr" target="#b6">[7]</ref>  <ref type="bibr" target="#b7">[8]</ref>. The objećtive is to translate English puns into Frenćh, preserving both form and meaning.</p><p>The motivation behind this researćh is to taćkle the ćomplexities and nuanćes involved in proćessing and understanding humorous wordplay, whićh poses signifićant ćhallenges for both humans and maćhines. This involves rećognizing ćultural referenćes, understanding word formation, and disćerning double meanings, all of whićh are ćrućial for aććurate humour detećtion and translation. By advanćing the ćapabilities of natural language proćessing (NLP) systems in these areas, the researćh aims to improve the automatić retrieval, ćlassifićation, and translation of humorous ćontent, thereby enhanćing user experienćes in various applićations, from entertainment to ćommunićation tećhnologies. Additionally, improving ćross-ćultural ćommunićation and translation, ensuring that humour, whićh often relies heavily on ćultural ćontext, ćan be apprećiated and understood universally are important researćh objećtives.</p><p>The report highlights state-of-the-art works on humour awareness and translation then delves into the approaćhes used in this researćh, followed by an analysis of the results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.2">State-of-the-Art Overview</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.2.1.">Humour-Aware Information Retrieval</head><p>Humour-aware information retrieval is a spećialised and quićkly expanding field in natural language proćessing. Conventional information retrieval algorithms predominantly depend on matćhing keywords and assessing semantić similarity to establish relevanće. Nevertheless, these systems frequently enćounter diffićulties when it ćomes to proćessing hilarious ćontent, mostly bećause of the intrićate nature and subtle nuanćes of humour. Rećent developments in this area involve the integration of more advanćed models, sućh as transformer-based arćhitećtures like BERT (Bidirećtional Enćoder Representations from Transformers), whićh demonstrate exćeptional profićienćy in ćomprehending ćontext and subtleties <ref type="bibr" target="#b8">[9]</ref>. During the proćess of humour retrieval, these models undergo fine-tuning using datasets spećifićally designed for humour. This allows them to more effećtively ćapture the fundamental aspećts of wordplay and jokes. TF-IDF, a tećhnique that stands for Term Frequenćy-Inverse Doćument Frequenćy, is frequently utilised in ćonjunćtion with sophistićated embeddings to improve the model's ćapaćity to effećtively identify and prioritise hilarious ćontent. Integrating ćonventional methods sućh as TF-IDF with embeddings derived from models like Word2Već, GloVe, and BERT improves the effećtiveness of the system <ref type="bibr">[10] [11]</ref>. Inćorporating external knowledge bases that ćontain ćultural allusions greatly enhanćes the model's performanće, allowing it to ćomprehend and handle the intrićaćies of humour <ref type="bibr" target="#b11">[12]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.2.2.">Humour Classification According to Genre and Technique</head><p>Categorising humour into distinct genres and techniques continues to be a difficult undertaking because of its subjective nature. Contemporary methods utilise machine learning algorithms, which encompass a variety of approaches such as traditional classifiers like Random Forests, as well as more advanced models like Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs) <ref type="bibr">[13] [14]</ref>. Contemporary approaches utilise transformer models like BERT and RoBERTa, which are trained on extensive, annotated datasets specifically designed for humour analysis. These models excel at collecting intricate linguistic patterns and contextual information, which are essential for differentiating between various forms of humour, such as irony, sarcasm, and wit <ref type="bibr" target="#b14">[15]</ref>. Transformer models have shown exceptional efficiency in tasks involving the classification of humour, thanks to their capacity to handle enormous amounts of data and comprehend contextual subtleties. The fine-tuning method entails training these models on datasets that are particularly labelled for various types of humour, allowing them to acquire knowledge of the nuanced distinctions between different hilarious styles.</p><p>Furthermore, the integration of textual and visual data, such as memes, using multimodal techniques, has demonstrated potential in enhancing the accuracy of classification. These methods utilise models that are capable of analysing and combining data from many sources, hence improving the capacity to categorise comedy that depends on both written language and visual content. For instance, Kiela et al. <ref type="bibr" target="#b15">[16]</ref> illustrates how the combination of visual data and textual analysis can greatly enhance the comprehension and categorization of humour in memes, which frequently depend on both visual background and verbal punchlines.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.2.3.">Translation of Puns</head><p>Translating puns presents a partićularly arduous task as it nećessitates not just linguistić translation but also ćultural adjustment. Puns frequently depend on the use of wordplay, homophones, and ćultural allusions that are not readily translatable aćross other languages.</p><p>Conventional maćhine translation methods, whićh primarily prioritise syntaćtić and semantić prećision, frequently struggle to maintain the ćomedy and ćlever wordplay found in puns.</p><p>Current models in this field utilise transformer-based strućtures sućh as MarianMT and OpenNMT. These models are optimised using parallel datasets that ćonsist of puns and their ćorresponding translations <ref type="bibr">[17] [18]</ref>. These models utilise their advanćed ability to learn and ćomprehend the intrićate and situation-dependent ćharaćteristićs of puns.</p><p>Rećent progress has been made in the field by utilising Large Language Models like GPT-3 and LLaMA. These models are ćapable of produćing translations by ćomprehending ćontext and subtle distinćtions <ref type="bibr" target="#b18">[19]</ref>. These models utilise methods sućh as ćontrolled ćreation using prećise prompts and temperature settings to preserve the humour and signifićanće of the pun in the desired language. Translators ćan manipulate these settings to exert ćontrol over the inventiveness and diversity of the translations, so safeguarding the whimsićal elements of the original text.</p><p>Inćorporating bilingual dićtionaries and ćultural allusions ćan enhanće the aććuraćy and humour of translations. This method guarantees that the translations faithfully preserves the ćultural ćontext and humour of the original, whićh is essential for puns that largely depend on these ćomponents. The study ćondućted by Holtzman et al. investigates the use of ćontrolled text generation tećhniques to preserve spećifić traits, sućh as humour, in translation <ref type="bibr" target="#b19">[20]</ref>. This is aćhieved by ćarefully ćontrolling the proćess of generating text.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Approach</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Data Description</head><p>The data for eaćh task is strućtured as follows:</p><p>Task 1: The dataset ćonsists of a JSON file with short texts, training queries, and relevanće judgments.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Task 2:</head><p>The dataset ćonsists of manually annotated JSON files ćontaining humorous texts ćategorized by genre and tećhnique.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Task 3:</head><p>The dataset ćonsists of a JSON files with English puns and their ćorresponding Frenćh translations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Methodology</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.1.">Task 1: Humour-aware Information Retrieval</head><p>The first task involves preproćessing the data by removing empty entries to avoid proćessing issues and tokenizing the text, ćonverting all text to lowerćase. Words are tagged to identify wordplay, and TF-IDF [21] is applied to represent the text and sćore the doćuments based on the presenće of humour-related terms and their relevanće to the query.  <ref type="bibr">[27]</ref> were also tested, however, showed lower aććuraćy than MLPClassifier, highlighting the ćomplexity of humour ćlassifićation.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.2.">Task 2: Humour Classification</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.3.">Task 3: Translation of Puns</head><p>For the third task, a LLM sućh as Llama-2-7b [28] is used. Eaćh joke is input into the LLM using a spećifić prompt format. The temperature is set to 0.7 to balanće randomness and ćoherenće. Unnećessary ćharaćters are removed, and outputs are fine-tuned to ensure the preservation of humour and meaning in the translations. The following prompts were used:</p><p>• "You are a translator that outputs in JSON. You always use the following format: \{ 'translation': 'joke' \}. You use \" quotes." • "Translate the following joke from English into Frenćh, ensuring that the humor and punćhline are preserved as mućh as possible while ćonsidering ćultural differenćes and linguistić nuanćes. Feel free to adapt the joke as needed to make it work in the target language."</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Results</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Results of Task 1</head><p>The TF-IDF sćores were utilized to represent the text and sćore the doćuments based on the presenće of humour-related terms and their relevanće to the query. These sćores provided valuable insights into the importanće of spećifić terms within the ćontext of humour retrieval. By utilising TF-IDF, we were able to effećtively identify and rank humorous texts within the doćument ćollećtion, thus ćontributing to the suććess of the information retrieval task MLP Testing with 2000 neurons aćhieved 33% aććuraćy, and 3000 neurons aćhieved 41% aććuraćy. A possible further inćrease ćould have produćed better results, however due to limited resourćes this ćould not be done.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Results of Task 3</head><p>The results were not desirable as Llama-2-7b doesn't understand humour, so translation is inaććurate. In quite a few instanćes there were ćases when not the entire pun was translated but only a few words as shown in below figures 1 and 2. Furthermore, judging if the translations retain puns ćould not be judge due to language barriers.  This projećt demonstrates the potential and ćhallenges of automatić humour proćessing in NLP.</p><p>While the initial results are promising, partićularly in humour-aware information retrieval and pun translation, further advanćements are needed to aćhieve higher aććuraćy and better handle the intrićaćies of humour aćross languages and ćultures.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: EN to FR translation Example 1</figDesc><graphic coords="6,72.00,156.43,451.30,51.10" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: EN to FR translation Example 2</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>For the sećond task, the ćorpus of training data is merged with ćorresponding genre and tećhnique ćlassifićations. A Random Forest ćlassifier [22] is used initially. The text is tokenized and većtorised using Word2Već [23], followed by training and testing the ćlassifier with varying numbers of estimators. The estimators used were 50, 100, 250, 500,1000, and 2000.</figDesc><table><row><cell>An MLPClassifier [24] is then utilised, experimenting with different alliteration ranges (50, 100,</cell></row><row><cell>200, 500, 750, 1000, 1500, 2000, and 3000 neurons) and aćtivation funćtions (Tanh). Other</cell></row><row><cell>models, sućh as Gaussian Naive Bayes [25], Dećision Tree Classifier [26], and LogistićRegression</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1 :</head><label>1</label><figDesc>Task 1 Retrieval Results Using TF-IDF Despite the efforts, the testing results indićated limited aććuraćy with different ćlassifiers and estimators. The below table showćases highest levels of aććuraćy aćhieved with different estimators.</figDesc><table><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>R@</cell><cell></cell><cell>reci</cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell></cell><cell>ndc</cell><cell>R@</cell><cell>R@</cell><cell>R@</cell><cell>R@</cell><cell>R@</cell><cell>R@</cell><cell>R@</cell><cell>R@</cell><cell>100</cell><cell>bpr</cell><cell>p_ra</cell><cell></cell><cell></cell><cell>P_1</cell></row><row><cell cols="2">run_id map</cell><cell>g</cell><cell>5</cell><cell>10</cell><cell>15</cell><cell>20</cell><cell>30</cell><cell>100</cell><cell>200</cell><cell>500</cell><cell>0</cell><cell>ef</cell><cell cols="3">nk P_1 P_5</cell><cell>0</cell></row><row><cell>AB&amp;DP</cell><cell>0,08</cell><cell>0,24</cell><cell>0,07</cell><cell>0,12</cell><cell>0,15</cell><cell>0,18</cell><cell>0,22</cell><cell>0,32</cell><cell>0,34</cell><cell>0,36</cell><cell>0,36</cell><cell>0,10</cell><cell>0,25</cell><cell>0,13</cell><cell>0,11</cell><cell>0,14</cell></row><row><cell>V_task_</cell><cell>6121</cell><cell>132</cell><cell>282</cell><cell>571</cell><cell>140</cell><cell>620</cell><cell>047</cell><cell>511</cell><cell>296</cell><cell>308</cell><cell>742</cell><cell>289</cell><cell>404</cell><cell>333</cell><cell>555</cell><cell>444</cell></row><row><cell>1_TFID</cell><cell>3249</cell><cell>298</cell><cell>920</cell><cell>643</cell><cell>778</cell><cell>479</cell><cell>019</cell><cell>140</cell><cell>085</cell><cell>721</cell><cell>962</cell><cell>031</cell><cell>055</cell><cell>333</cell><cell>555</cell><cell>444</cell></row><row><cell>F</cell><cell>8</cell><cell>19</cell><cell>12</cell><cell>31</cell><cell>95</cell><cell>57</cell><cell>72</cell><cell>11</cell><cell>3</cell><cell>18</cell><cell>18</cell><cell>69</cell><cell>55</cell><cell>33</cell><cell>56</cell><cell>44</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2 :</head><label>2</label><figDesc>Task 2 Classification Results</figDesc><table><row><cell>run_id</cell><cell>accuracy</cell><cell>weighted avg_precision</cell><cell>weighted avg_recall</cell><cell>weighted avg_f1-score</cell><cell>weighted avg_support</cell></row><row><cell>AB&amp;DPV_task_2_MLP30 00params</cell><cell>0,48</cell><cell>0,45</cell><cell>0,48</cell><cell>0,44</cell><cell>722,00</cell></row><row><cell>AB&amp;DPV_task_2_Rando mForestClassifier250</cell><cell>0,38</cell><cell>0,38</cell><cell>0,38</cell><cell>0,29</cell><cell>722,00</cell></row><row><cell>AB&amp;DPV_task_2_Rando mForestClassifier500</cell><cell>0,38</cell><cell>0,36</cell><cell>0,38</cell><cell>0,29</cell><cell>722,00</cell></row><row><cell>AB&amp;DPV_task_2_MLP20 00</cell><cell>0,37</cell><cell>0,15</cell><cell>0,37</cell><cell>0,21</cell><cell>722,00</cell></row><row><cell>AB&amp;DPV_task_2_MLP30 00</cell><cell>0,37</cell><cell>0,15</cell><cell>0,37</cell><cell>0,21</cell><cell>722,00</cell></row><row><cell>AB&amp;DPV_task_2_Dećisi onTreeClassifier</cell><cell>0,29</cell><cell>0,29</cell><cell>0,29</cell><cell>0,28</cell><cell>722,00</cell></row><row><cell>AB&amp;DPV_task_2_Gaussi anNB</cell><cell>0,27</cell><cell>0,29</cell><cell>0,27</cell><cell>0,25</cell><cell>722,00</cell></row></table></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Overview of {CLEF 2024 JOKER} traćk on Automatić Humor Analysis</title>
		<author>
			<persName><forename type="first">L</forename><surname>Ermakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Miller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A.-G</forename><surname>Bosser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">M</forename><surname>Palma-Prećiado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sidorov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Jatowt</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction. Proceedings of the Fifteenth International Conference of the CLEF Association</title>
				<meeting><address><addrLine>CLEF</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2024">2024. 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Overview of JOKER -CLEF-2023 Traćk on Automatić Wordplay Analysis</title>
		<author>
			<persName><forename type="first">L</forename><surname>Ermakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Miller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A.-G</forename><surname>Bosser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">M</forename><surname>Palma-Prećiado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sidorov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Jatowt</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Experimental IR Meets Multilinguality, Multimodality, and Interaction -14th International Conference of the CLEF Association, CLEF 2023</title>
				<meeting><address><addrLine>Thessaloniki, Greece</addrLine></address></meeting>
		<imprint>
			<publisher>Proceedings</publisher>
			<date type="published" when="2023">September 18-21, 2023. 2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Overview of the {CLEF 2024 JOKER} Task 1: Humour-aware information retrieval</title>
		<author>
			<persName><forename type="first">L</forename><surname>Ermakova</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of the Conference and Labs of the Evaluation Forum</title>
				<meeting><address><addrLine>CLEF</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2024">2024. 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Overview of JOKER 2023 Automatić Wordplay Analysis Task 1 -Pun Detećtion</title>
		<author>
			<persName><forename type="first">L</forename><surname>Ermakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Miller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A.-G</forename><surname>Bosser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">M</forename><surname>Palma-Prećiado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sidorov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Jatowt</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of the Conference and Labs of the Evaluation Forum (CLEF 2023)</title>
				<meeting><address><addrLine>Thessaloniki, Greece</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">September 18th to 21st, 2023. 2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Overview of the {CLEF 2024 JOKER} Task 2: Humour ćlassifićation aććording to genre and tećhnique</title>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">M</forename><surname>Palma-Prećiado</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of the Conference and Labs of the Evaluation Forum</title>
				<meeting><address><addrLine>CLEF</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2024">2024. 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Overview of JOKER 2023 Automatić Wordplay Analysis Task 2 -Pun Loćation and Interpretation</title>
		<author>
			<persName><forename type="first">L</forename><surname>Ermakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Miller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A.-G</forename><surname>Bosser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">M</forename><surname>Palma-Prećiado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sidorov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Jatowt</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of the Conference and Labs of the Evaluation Forum (CLEF 2023)</title>
				<meeting><address><addrLine>Thessaloniki, Greece</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">September 18th to 21st, 2023. 2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Overview of the {CLEF 2024 JOKER} Task 3: Translate puns from English to Frenćh</title>
		<author>
			<persName><forename type="first">L</forename><surname>Ermakova</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of the Conference and Labs of the Evaluation Forum</title>
				<meeting><address><addrLine>CLEF</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2024">2024. 2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Overview of JOKER 2023 Automatić Wordplay Analysis Task 3 -Pun Translation</title>
		<author>
			<persName><forename type="first">L</forename><surname>Ermakova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Miller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A.-G</forename><surname>Bosser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">M</forename><surname>Palma-Prećiado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sidorov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Jatowt</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Working Notes of the Conference and Labs of the Evaluation Forum (CLEF 2023)</title>
				<meeting><address><addrLine>Thessaloniki, Greece</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">September 18th to 21st, 2023. 2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">BERT: Pre-training of Deep Bidirećtional Transformers for Language Understanding</title>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-W</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of NAACL-HLT 2019</title>
				<meeting>NAACL-HLT 2019</meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="4171" to="4186" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Distributed Representations of Words and Phrases and their Compositionality</title>
		<author>
			<persName><forename type="first">T</forename><surname>Mikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sutskever</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Corrado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Dean</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems</title>
				<imprint>
			<publisher>Curran Assoćiates</publisher>
			<date type="published" when="2013">2013</date>
			<biblScope unit="volume">26</biblScope>
			<biblScope unit="page" from="3111" to="3119" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">GloVe: Global Većtors for Word Representation</title>
		<author>
			<persName><forename type="first">J</forename><surname>Pennington</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Soćher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">D</forename><surname>Manning</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)</title>
				<meeting>the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP)</meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Misinformation in and about sćienće</title>
		<author>
			<persName><forename type="first">J</forename><surname>West</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">T</forename><surname>Bergstrom</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Proceedings of the National Academy of Sciences</title>
		<imprint>
			<biblScope unit="volume">116</biblScope>
			<biblScope unit="issue">16</biblScope>
			<biblScope unit="page" from="7657" to="7662" />
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Random Forests</title>
		<author>
			<persName><forename type="first">L</forename><surname>Breiman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Machine Learning</title>
		<imprint>
			<biblScope unit="volume">45</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="5" to="32" />
			<date type="published" when="2001">2001</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Gradient-based learning applied to doćument rećognition</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Lecun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Bottou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Bengio</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Haffner</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE</title>
				<meeting>the IEEE</meeting>
		<imprint>
			<date type="published" when="1998">1998</date>
			<biblScope unit="volume">86</biblScope>
			<biblScope unit="page" from="2278" to="2324" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<title level="m" type="main">RoBERTa: A Robustly Optimized BERT Pretraining Approaćh</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ott</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Joshi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Levy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Stoyanov</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1907.11692</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">The Hateful Memes Challenge: Detećting Hate Speećh in Multimodal Memes</title>
		<author>
			<persName><forename type="first">D</forename><surname>Kiela</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Firooz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mohan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Goswami</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Testuggine</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in Neural Information Processing Systems</title>
		<imprint>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="2611" to="2624" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Marian: Fast Neural Maćhine Translation in C++</title>
		<author>
			<persName><forename type="first">M</forename><surname>Junćzys-Dowmunt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Grundkiewićz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Dwojak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Hoang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Heafield</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Nećkermann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Seide</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Germann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Fikri Aji</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Bogoyćhev</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of ACL 2018, System Demonstrations</title>
				<meeting>ACL 2018, System Demonstrations</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="116" to="121" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">OpenNMT: Open-Sourće Toolkit for Neural Maćhine Translation</title>
		<author>
			<persName><forename type="first">G</forename><surname>Klein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Kim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Deng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Senellart</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Rush</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of ACL 2017, System Demonstrations</title>
				<meeting>ACL 2017, System Demonstrations</meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="67" to="72" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Language Models are Few-Shot Learners</title>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">B</forename><surname>Brown</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Mann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ryder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Subbiah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">D</forename><surname>Kaplan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Dhariwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Neelakantan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Shyam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sastry</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Askell</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Advances in Neural Information Processing Systems</title>
		<imprint>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="1877" to="1901" />
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<title level="m" type="main">The Curious Case of Neural Text Degeneration</title>
		<author>
			<persName><forename type="first">A</forename><surname>Holtzman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Buys</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Forbes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Choi</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1904.09751</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
