<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">On Cross-Language Entity Label Projection and Recognition</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Paolo</forename><surname>Gajo</surname></persName>
							<email>paolo.gajo2@unibo.it</email>
							<affiliation key="aff0">
								<orgName type="institution">Università di Bologna</orgName>
								<address>
									<addrLine>Corso della Repubblica, 136</addrLine>
									<postCode>47121</postCode>
									<settlement>Forlì</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Alberto</forename><surname>Barrón-Cedeño</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">Università di Bologna</orgName>
								<address>
									<addrLine>Corso della Repubblica, 136</addrLine>
									<postCode>47121</postCode>
									<settlement>Forlì</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">On Cross-Language Entity Label Projection and Recognition</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">64173FD76F35DD8B2CD55EAE37D8A636</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T17:36+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>information extraction</term>
					<term>named entity recognition</term>
					<term>cross-lingual label projection</term>
					<term>data augmentation</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Most work on named entity recognition (NER) focuses solely on English. Through the use of training data augmentation via machine translation (MT), multilingual NER can become a powerful tool for information extraction in multilingual contexts. In this paper, we augment NER data from culinary recipe ingredient lists by means of MT and word alignment (WA), following two approaches: (i) translating each entity separately, while taking into account the full context of the list and (ii) translating the whole list of ingredients and then aligning entities using three types of WA models: Giza++, Fast Align, and BERT, fine-tuned using a novel entity-shuffling approach. We depart from English data and produce Italian versions via MT, span-annotated with the entities projected from English. Then, we use the data produced by the two approaches to train mono-and multilingual NER BERT models. We test the performance of the WA and NER models on an annotated dataset of ingredient lists, partially out-of-domain compared to the training data. The results show that shuffling entities leads to better BERT aligner models. The higher quality NER data created by these models enables NER models to achieve better results, with multilingual models reaching performances equal to or greater than their monolingual counterparts.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Named entity recognition (NER) is a sequence labeling task with a long history of works mainly focusing on the recognition of entities such as people, locations, and organizations. Multilingual NER has also attracted research efforts, with recent SemEval campaigns including tasks on multilingual complex NER (MultiCoNER) <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2]</ref>. Despite its popularity and various mono-and multilingual NER datasets being available, specific domains such as the culinary one likely require new annotated data. In addition, NER is often the first step in information extraction for knowledge graph construction and, to the best of our knowledge, all literature in the domain of cuisine on this topic solely focuses on English data <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4,</ref><ref type="bibr" target="#b4">5,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b6">7]</ref>. Therefore we argue that, given cuisine's multicultural nature, more research in this direction is warranted.</p><p>Entity label projection <ref type="bibr" target="#b7">[8]</ref> aims to address this scarcity by automating the data generation process for NER. This task consists in taking the labels associated with spans from a source text and automatically applying them to its translation in another language, i.e. the target text. Through this task, we attempt to find an efficient automatic way of developing models for entity projection across languages to produce high-quality multilingual data for recipe Named Entities (r-NE) <ref type="bibr" target="#b3">[4]</ref>. Departing from an English-language dataset containing ingredients from culinary recipes, annotated at the span level with entity category labels, we first rely on a MT engine to translate each source entity 𝑠𝑖 individually into Italian, while keeping the full context into account. This results in a first entity-wise (EW) translated EN-IT-ES dataset where entities are linked across languages. 1  Using these synthetic alignments, we train BERT models to align source and target entities, shuffling the latter to prevent the model from learning to simply predict the original entity order. We then test the models on two novel entity alignment datasets, partially out-of-domain compared to the training data, e.g., as regards the used food products, units of measure, and cooking processes. As baselines to evaluate the BERT alignment models, we use Giza++ <ref type="bibr" target="#b8">[9]</ref> and Fast Align <ref type="bibr" target="#b9">[10]</ref>, two statistical word alignment (WA) models. In order to produce higherquality r-NE data, we translate the ingredient lists across their whole length, predicting target entity spans with the best BERT models from the previous step, along with the baseline models. We thus obtain various sentencewise (SW) translated datasets in Italian, trading some alignment accuracy for better translations.</p><p>Both types of training data, EW and SW, are then used to fine-tune mono-and multilingual BERT NER models on the task of recognizing entities in food recipes. The models are trained on various combinations of mono-and multilingual data and are tested on the entity annotations from the two aforementioned novel testing datasets.</p><p>Our contribution is three-fold: (i) We show the efficacy of fine-tuning alignment models by shuffling entities in contexts where most of the information depends on the presence of lexical items rather than the dependencies linking them. (ii) We showcase the performance delta between mono-and multilingual NER models when finetuning on the synthetic data produced by our alignment. These models can be used to label large datasets in multiple languages at a finer granularity level compared to currently available monolingual resources. (iii) We release code and data to produce data in multiple languages. <ref type="foot" target="#foot_0">2</ref>The rest of the paper is structured as follows. Section 2 presents relevant past research on the subjects of cross-lingual entity alignment and recognition. Section 3 introduces the datasets and corpora used in the experiments, along with their annotation process. Section 4 presents architecture, training, and evaluation details for the models comprising our pipeline. Section 5 discusses the conducted experiments and their results. Finally, Section 6 summarizes the paper and draws conclusions. Appendix A shows further results including Spanish. Appendix B presents statistics and gives insight on the additional training data used. Appendix C lists information on the computational requirements.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head><p>Word alignment was first approached for statistical MT, with models such as IBM 1-5 <ref type="bibr" target="#b10">[11]</ref>, used in well-known implementations such as Giza++ and Fast Align. With the advent of Transformers <ref type="bibr" target="#b11">[12]</ref> and the BERT model <ref type="bibr" target="#b12">[13]</ref>, this task has been approached by employing both question answering <ref type="bibr" target="#b13">[14]</ref> and token classification <ref type="bibr" target="#b14">[15]</ref> models, trained on freely available resources, such as XL-WA <ref type="bibr" target="#b15">[16]</ref>.</p><p>A number of past works have studied label projection following a range of approaches. Jain et al. <ref type="bibr" target="#b7">[8]</ref> project PER, ORG, LOC and MISC labels (person, organization, location, and miscellaneous) by translating sentences and then finding potential matches using glossaries. Fei et al. <ref type="bibr" target="#b16">[17]</ref> align words using Fast Align and use POS tagging to enhance data for semantic role labeling. García-Ferrero et al. <ref type="bibr" target="#b17">[18]</ref> use the AWESoME word alignment model <ref type="bibr" target="#b19">[19]</ref> to align machine-translated data from NER datasets in seven languages. Li et al. <ref type="bibr" target="#b14">[15]</ref> fine-tune a NER model on English PER, ORG, LOC, MISC data from CoNLL2003 <ref type="bibr" target="#b20">[20]</ref> to infer on the source portion of parallel Opus corpora <ref type="bibr" target="#b21">[21]</ref> with the aim of creating silver NER data. Subsequently, they train an XLM RoBERTa alignment model by using Wikipedia articles and project the labels on the target portion of the parallel corpus, which they use to train a target-language NER model.</p><p>NER can also be approached with large language models (LLM) <ref type="bibr" target="#b22">[22,</ref><ref type="bibr" target="#b23">23,</ref><ref type="bibr" target="#b24">24]</ref> by prompting them to extract entities from a given text. For example, PromptNER <ref type="bibr" target="#b25">[25]</ref> uses chain of thought <ref type="bibr" target="#b26">[26]</ref> along with a list of entity definitions to prompt a variety of LLMs, obtaining results on par with SOTA supervised NER systems. Similarly, <ref type="bibr" target="#b27">[27]</ref> use in-context learning <ref type="bibr" target="#b28">[28]</ref> to evaluate GPT-3 <ref type="bibr" target="#b22">[22]</ref> for NER on the CoNLL2003 <ref type="bibr" target="#b20">[20]</ref> and OntoNotes5.0 <ref type="bibr" target="#b29">[29]</ref> datasets by using retrieval-augmented generation <ref type="bibr" target="#b30">[30]</ref> and comparing the results to BERT and models based on graph neural networks <ref type="bibr" target="#b31">[31]</ref>.</p><p>With regard to data specific to the culinary domain, many English-language resources exist in various forms. RecipeDB <ref type="bibr" target="#b32">[32]</ref> is an ontology comprising 118 𝑘 web recipes which can be used to relate foods and cooking processes to taste profiles and health data. FoodOn <ref type="bibr" target="#b33">[33]</ref> is a "farm-to-fork" ontology which provides a structure of relationships between food products across the whole industrial supply chain. Bridging the gap between ontologies and NER datasets, FoodKG <ref type="bibr" target="#b34">[34]</ref> is a knowledge graph which can be used to find ingredient substitutions based on dietary health requirements. It is built by leveraging FoodOn and Recipe1M+ <ref type="bibr" target="#b35">[35]</ref>, a dataset originally intended for learning joint text/image embeddings on over 1 𝑀 culinary recipes. Expanding on Recipe1M+, Bień et al. <ref type="bibr" target="#b36">[36]</ref> construct RecipeNLG, comprising more than 2 𝑀 recipes. It is the biggest food NER dataset to date, but its granularity stops at the sole food product names. More fine-grained silver labels are obtained by Komariah et al. <ref type="bibr" target="#b37">[37]</ref>, who propose a new methodology to extract entities from AllRecipes. <ref type="foot" target="#foot_1">3</ref> Doing so, they construct FINER, a dataset comprising 64 𝑘 recipes with labels predicted by what the authors refer to as a "semi-supervised multi-model prediction technique. " The dataset also contains recipe tags such as vegetarian and vegan, which can be useful for training recipe classifiers. Leveraging RecipeDB <ref type="bibr" target="#b32">[32]</ref>, a large-scale structured corpus of recipes, <ref type="bibr" target="#b38">[38]</ref> generate a synthetic dataset of augmented ingredient phrases and compare the NER performance of various rule-based and neural models.</p><p>Despite the wide availability of English-language resources in the culinary domain, other languages are largely understudied. To the best of our knowledge, the only study to approach this domain in a multilingual setting was conducted by Radu et al. <ref type="bibr" target="#b39">[39]</ref>, who obtain NER tags automatically in English, German, and French by using a regex-based tagger. Our work aims to partially address this gap in past research by focusing on Italian.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Data</head><p>The entity alignment data used for training is generated through MT starting from TASTEset <ref type="bibr" target="#b40">[40]</ref>, a dataset comprising ingredient lists from 700 food recipes, annotated at the span level. We use TASTEset because it is human-curated and its annotations are fine-grained. We translate each entity one by one with DeepL, <ref type="foot" target="#foot_2">4</ref> concurrently feeding the whole ingredient list and the single entity as two separate inputs. This provides DeepL with context, improving translation quality and retaining the start and end span indexes in the target text by simply concatenating each translated entity. To the best of our knowledge, DeepL is currently the only MT engine capable of contextually translating a substring taken from a sentence, which is why we are using it in this study. Doing this, we obtain an Entity-wise Machine-translated TASTEset (EMT). Since entities are automatically paired to the source label, the distribution across English and Italian is identical (Table <ref type="table" target="#tab_0">1</ref>).</p><p>We also generate shuffled variations of EMT, where the entities within a single ingredient have a probability 𝑝 ∈ {0.1, 0.2, . . . , 1.0} of being shuffled, for a total of ten variations. Figure <ref type="figure">1</ref> shows an example where entities have been shuffled in the first and third target ingredients. The rationale behind this approach is that, when training on EMT, if the dataset were to be left as-is, the model would simply learn to associate a source entity to the target entity in the corresponding position, since entities are simply translated and replaced in EMT.</p><p>Overall, we have 22 different variations of EMT, i.e. the original and the 10 shuffled versions for each of the two types of tokenization (mBERT's WordPiece <ref type="bibr" target="#b12">[13]</ref> vs mDeBERTa's SentencePiece <ref type="bibr" target="#b41">[41]</ref>). The datasets have to be tokenized during the generation of the dataset because token indexes depend on the tokenizer being used when converted from character-level span annotations.</p><p>We produce a second kind of synthetic dataset by first translating the ingredient lists as a whole, and then aligning source and target entities by using the BERT, Giza++, and Fast Align models presented in Section 4. We refer to this type of dataset as Sentence-wise Machine-translated TASTEset (SMT). As Table <ref type="table" target="#tab_0">1</ref> shows, the SMT dataset produced by the BERT model trained on both XL-WA and the shuffled version of EMT contains slightly fewer entities than the source material. This is due to the fact that at times the models produce impossible predictions, e.g. predicting the end of an entity to be before its start. <ref type="foot" target="#foot_3">5</ref> This problem does not exist with Giza++ and Fast Align, since their alignments are word-based. As additional training data for the BERT models, we use the EN-IT portion of XL-WA. Table <ref type="table" target="#tab_10">9</ref> in Appendix B reports the size of each of the partitions we used.</p><p>For testing, we annotated an English-Italian dataset of recipes, obtained from GialloZafferano (where the English recipes are translated from the Italian ones). <ref type="foot" target="#foot_4">6</ref> For the annotation process, we recruited a professional translator who is a native speaker of Italian, with an MA in Specialized Translation in both English and Spanish. Figure <ref type="figure" target="#fig_0">2</ref> shows the instructions given for the first multi-class entity annotation task, which consider the same entities as</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Class</head><p>EMT (en/it) SMT (it) GZ (en) GZ(it) food 4,020 TASTEset, and the second cross-language entity-linking annotation task, carried out by the same annotator at a later time. The annotation was carried out in Label Studio. <ref type="foot" target="#foot_5">7</ref>The GialloZafferano (GZ) dataset comprises 597 recipes. The alignments were annotated manually on a subset of 300 recipes, with the possibility of more than one source entity being aligned with one target entity, and vice versa. This is because some recipes contain more than one ingredient option in English but not in Italian (and vice versa), e.g., Cocomero (anguria) 1 fetta vs Watermelon 1 slice. The GZ dataset contains a total of 46,903 NER annotations and 9,842 alignments.</p><p>We manually scrutinized GZ and found that the paired recipes do not always coincide completely. Some ingredients may be missing in either language or be an equivalent rather than the same food product. In order to avoid training the alignment models on excessively different recipes, we chose to avoid annotating alignments whenever the number of source ingredients missing from the target recipe surpassed a heuristic threshold of 1/3.</p><p>Note that in GZ quantities and units of measure are localized and are thus listed in both imperial and SI units. As shown in Table <ref type="table" target="#tab_0">1</ref>, this is reflected by the lower number of instances annotated as quantity and unit in the Italian portion of GZ, compared to its English portion.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Instructions for the multi-class entity annotation task.</head><p>Annotate the ingredients below by assigning to spans of text one of the following categories: quantity, unit, food, color, part, physical quality, process, purpose, taste. Use quantity for numerical values or expressions such as 'to taste', identifying the quantity of an ingredient. Unit stands for "unit of measure", such as grams (g) or ounces (oz). Use color for any color that is not part of a food's own name (e.g. 'red' can be tagged in 'red wine', but not 'black' in 'blackberries'). Part refers to parts of an ingredient, such as 'wings' in 'chicken wings'. Use physical quality for attributes which already characterize an ingredient at the start of the preparation. Process refers to actions that the reader is supposed to carry out. The label purpose answers the question, "What is this entity used for?" Finally, use taste for words referring to a taste, such as 'unsweetened' or 'dry', with relation to a wine. N.B.: The annotations cannot overlap. You can either choose to annotate multiple spans with the same label, or avoid annotating one or more spans of text.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Instructions for the cross-lingual entity linking task.</head><p>Link each source language entity to its corresponding entity in the target language with an arrow. Entities should only be linked if they share the same use. For example, the "2" in "2 tablespoons chopped onions" does not have the same function as in "2 cebollas largas picadas", since one refers to tablespoons and other to the number of onions. However, "onions" and "chopped" could still be linked, as they are equivalent in the two sequences. Individual source entities can be linked to multiple target entities and vice versa. N.B.: Entities can still be linked if they differ slightly in form or content but still clearly perform the same function in the same ingredient context. For example, "340" and "450" could still be linked if they both refer to the quantity of grams of the same source and target food products. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Models</head><p>Entity Alignment As baselines, we use two statistical models: Giza++ <ref type="bibr" target="#b8">[9]</ref> and Fast Align <ref type="bibr" target="#b9">[10]</ref>. Giza++ combines a HMM <ref type="bibr" target="#b42">[42]</ref> alignment model and IBM M1-5 <ref type="bibr" target="#b10">[11]</ref>. Fast Align is much more lightweight, only leveraging IBM M2. We use two multilingual BERT models as well: mBERT <ref type="bibr" target="#b12">[13]</ref> as the baseline multilingual Transformer model and mDeBERTa <ref type="bibr" target="#b43">[43]</ref> because of its larger size (276𝑀 vs 179𝑀 param.) and performance. When using the BERT models, we follow Nagata et al. <ref type="bibr" target="#b13">[14]</ref> and treat entity alignment as a question-answering task, enclosing the source word to be aligned within rarely used characters, e.g., '•', feeding the model both the source sequence 𝐴 and the target sequence 𝐵 at once. Figure <ref type="figure">1</ref>  an entity within a shuffled ingredient's boundaries. We train the models for up to 3 epochs on each dataset with a batch size of 16. The optimizer's learning rate is set at 3 × 10 −4 , while 𝜖 is 10 −8 . Each training run, we select the best model based on the Exact metric 𝐸 <ref type="bibr" target="#b44">[44]</ref>:</p><formula xml:id="formula_0">𝐸 = ∑︀ 𝑛 𝑖 𝑒𝑥𝑎𝑐𝑡(𝑝𝑖, 𝑔𝑖) ‖𝑝𝑟𝑒𝑑𝑠‖ ,<label>(1)</label></formula><p>where 𝑝𝑟𝑒𝑑𝑠 is a list of predictions and 𝑒𝑥𝑎𝑐𝑡(𝑝𝑖, 𝑔𝑖) is the Kronecker delta:</p><formula xml:id="formula_1">𝑒𝑥𝑎𝑐𝑡(𝑝𝑖, 𝑔𝑖) = {︃ 1, if 𝑝𝑖 = 𝑔𝑖, 0, if 𝑝𝑖 ̸ = 𝑔𝑖<label>(2)</label></formula><p>with the predicted and gold strings 𝑝𝑖 and 𝑔𝑖 having been lowercased and stripped of excess punctuation and spaces. We calculate mean Exact and its standard deviation out of five random runs for each model. In order to improve the models' ability to align entities, we optionally train them on an intermediary wordalignment task using the EN-IT training and dev sets of XL-WA. In addition, we train mBERT and mDeBERTa solely using said XL-WA partitions in order to test them directly on GZ. This serves as a baseline which will allow us to gauge the positive effects of fine-tuning on EMT. Entity Recognition For the NER task, treated as token classification, we once again use mBERT. <ref type="foot" target="#foot_6">8</ref>To test the efficacy of the multilingual approach, we also use the following monolingual models when training and testing on a single language: bert-base-uncased (henceforth "BERTen") for English <ref type="bibr" target="#b12">[13]</ref> and bert-base-italian-uncased ("BERT it ") <ref type="bibr" target="#b45">[45]</ref> for Italian. We forgo mDeBERTa for this task, as the focus is showing a comparison between models of equivalent size and performance. Prior to training, the data is preprocessed and labeled using the BIO annotation scheme <ref type="bibr" target="#b46">[46]</ref>. We ignore subword tokens when calculating cross-entropy loss, following established methodology. <ref type="foot" target="#foot_7">9</ref>We train the models on the EN-IT, EN-ES, EN-IT-ES language subsets of EMT and of the four versions of SMT, produced by mBERT, mDeBERTa, Giza++, and Fast Align. For the BERT models, we use the same hyperparameters used for the alignment task, but with a lower learning rate of 2 × 10 −4 . The models are evaluated using the macro F1-measure. Details on the employed computational resources can be found in Appendix C.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Class Fast Align</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Results and Discussion</head><p>Entity Alignment Table <ref type="table" target="#tab_1">2</ref> reports the Exact scores for the entity alignment experiment. The entity shuffling approach appears to be very effective for creating data which can make the models better at generalizing. The performance of every single model is greatly enhanced when shuffling ingredients just 10% of the time, with increased shuffling frequency not leading to any significant further improvement. The increase in performance seems to be greater for models which have undergone intermediate training on XL-WA, with mDeBERTax gaining almost 12 points in the Exact metric, when fine-tuned on shuffled data. Unsurprisingly, the larger mDeBERTa performs much better than the smaller mBERT across the board. Although the model obtaining the highest mean performance is obtained at 𝑃 = 0.8, an overlap can be observed between all the confidence intervals for 𝑃 ≥ 0.1. However, this is not true when going from 𝑃 = 0 to 𝑃 = 0.1. Consequently, increased shuffling past 10% does not seem to provide a concrete performance gain, which is why we decided to produce SMT by using the BERT trained on the least-shuffled version of EMT.</p><p>In and of itself, the intermediary training step on XL-WA provides a slight performance boost when looking at mBERT vs mBERTx and mDeBERTa vs mDeBERTax. Still, this increase is much smaller compared to the one gained through shuffling. While finetuning the models on a general word-alignment task can be beneficial, the target domain is likely too different from the training data for this to produce a large performance boost. This is especially true as regards the structure of the sentences, since the test data is comprised by short lists of entities separated by semicolons, while the training data is a domain-balanced sample of sentences from Wikipedia. An additional performance boost is provided by multilingual fine-tuning, while cross-lingual settings (e.g., fine-tuning on ES and testing on IT) lead to worse outcomes. Table <ref type="table" target="#tab_6">6</ref> (Appendix A) shows the results.</p><p>Table <ref type="table" target="#tab_3">3</ref> reports the performance of the best overall models on each class. As the results show, the much lighter Giza++ model surpasses mBERTx, only trailing behind mDeBERTax. The poor scores achieved by the two BERT models are largely attributable to their poor scores on the unit and part classes. We hypothesize that this poor class-specific performance has to do with units of measure often being very short strings. Training mDeBERTa only on the unit instances does not improve its performance, with the model scoring a lower 18.08 Exact metric. Inspecting its individual predictions in this single-class scenario, we noticed that the model does learn to always predict two consecutive tokens, but the enclosed token does not match the original text when converted into characters. This is due to two separate issues: (i) the model selects the wrong span, e.g., selecting an ingredient such as "carote" (carrots) rather than the unit "g" or (ii) the model's prediction is empty when converted to characters. Since mBERT and mDeBERTa both have poor performance on this class while using two different tokenization algorithms (WordPiece vs Senten-cePiece), the problem may lie in the models' tokenizer's token-to-character conversion method. <ref type="foot" target="#foot_8">10</ref> We plan to shed light on this in the future. As regards the part class, the poor performance could be explained by the small More generally, Fast Align, being the simplest and most lightweight model, performs on average well below the other more complex models. <ref type="table" target="#tab_4">4</ref> reports the results for the NER task. The aligner column indicates which alignment model, out of the best ones listed in Table <ref type="table" target="#tab_3">3</ref>, has produced the SMT training data used to fine-tune the NER model. When no alignment model is specified, the training data being used is EMT. Note that in this case we are not using EMT's shuffled versions, as there is no relation between any two recipes when fine-tuning on the NER task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Entity Recognition Table</head><p>When training and testing on Italian data, the best results are obtained for both mBERT and BERT it when fine-tuning on SMT data produced by mDeBERTa. When fine-tuning them on EMT, the performance is noticeably lower, with a 5-point difference for mBERT and an 8-point difference for BERT it . The data produced by mBERT also allows both models to outperform the EMT baseline, although by smaller amounts. Conversely, the data produced by Fast Align and Giza++ worsens the data quality in 75% of the cases. When fine-tuning mBERT on bilingual ES-IT data, the performance on the test set remains essentially unvaried (see Table <ref type="table" target="#tab_9">8</ref> in Appendix A).</p><p>Looking at the baselines at the bottom of Table <ref type="table" target="#tab_4">4</ref>, we can see that fine-tuning mBERT on English data yields worse performance when testing on GZ, compared to fine-tuning on EMT's Italian data. Our data augmentation strategy is thus providing an evident performance boost, with entity alignment producing bigger improvements than machine-translating each entity individually.</p><p>In all settings, mBERT performs on par with the monolingual models. This shows that a single multilingual model can suffice when extracting entities from multilingual corpora, saving time and compute.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusions</head><p>We explored a simple novel technique to automatically generate high-quality multilingual NER data by combining machine translation and cross-language entity linking. For our experiments, we relied on the Englishlanguage TASTEset dataset, which includes recipes whose lists of ingredients are span-annotated for entity recognition. Moreover, we manually curated a novel English-Italian cross-language dataset, featuring the same kind of annotation, with the addition of crosslanguage alignments.</p><p>We machine translated the entities in TASTEset's recipes individually and shuffled them within ingredient boundaries. Leveraging this augmented data, we then fine-tuned BERT entity-alignment models. Using statistical word-alignment models as baselines, we tested these BERT models on our English-Italian parallel corpus. The results showed that models fine-tuned using our novel approach consistently outperform those trained on unshuffled data, along with two statistical baselines.</p><p>We then created additional synthetic data by first translating TASTEset's recipes in their entirety, and then aligning the entities in the machine-translated target text using the best models obtained from the first part of the study. These data allowed us to obtain better NER models, compared to the ones we would have obtained by using the original recipes translated entity by entity. We tested monolingual English and Italian BERT models against mBERT, and showed that the latter is capable of obtaining the same performance as its monolingual counterparts when tested on monolingual NER data.</p><p>In future work, we plan to extend the annotation of our datasets, both in terms of number of instances and annotators. We will also prioritize solving the token-tocharacter conversion issues encountered in this study. Furthermore, we plan to leverage this data augmentation technique in order to improve multilingual text-to-graph models, since all of the literature in this regard focuses on English-only data <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b3">4,</ref><ref type="bibr" target="#b4">5,</ref><ref type="bibr" target="#b5">6,</ref><ref type="bibr" target="#b6">7]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A. Incorporating Spanish</head><p>In order to test more thoroughly the soundness of our approach, we carry out an equivalent study with Spanish.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A.1. Data</head><p>We annotated an English-Spanish dataset of recipes obtained from My Colombian Recipes, <ref type="foot" target="#foot_9">11</ref> which we refer to as MCR. MCR is translated from English to Spanish, which is evident from the fact that on the website all Spanish recipes have an English counterpart, but not vice versa. We believe approximately 5-10% of the dataset's instances to be possible MT. A good indication of this is the fact that the English "to taste" is sometimes translated as "para probar", likely an MT mistake, while other times the correct "al gusto" is used. Although using machinetranslated data is not ideal, this was our best choice for a Spanish-language parallel recipe corpus, due to the lack of availability of similar online resources. The use of MT data has implications with respect to the evaluation of the models, as their performance would likely be lower in a real-world scenario involving recipes written directly in Spanish. Nonetheless, given the limited amount of data we hypothesize as being machine-translated, we believe the impact would not be large enough to discredit our results, which focus on the improvement over the cross-lingual EN-ES baseline, rather than the absolute performance of the best model.</p><p>MCR contains 276 recipes, 104 of which are bilingual and annotated with alignments. Due to this imbalance between the number of English and Spanish recipes, the number of entities is around 3x for the former, as shown in Table <ref type="table" target="#tab_5">5</ref>. In total, MCR contains annotations for 15,257 entities and 3,565 alignments. Along with the ingredient lists, MCR also contains cooking instructions for all its recipes, along with nutritional facts for 139 of them.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A.2. BERT Model</head><p>As a monolingual Spanish BERT model baseline to compare against mBERT, we use bert-base-spanish-wwm-cased ("BERTes") <ref type="bibr" target="#b47">[47]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A.3. Results</head><p>Entity Alignment Table <ref type="table" target="#tab_6">6</ref> reports the results for the alignment task, complete with the settings including Spanish-language data.</p><p>Fine-tuning on the same language as the test set yields better results than cross-lingual scenarios. Furthermore, the best performance on MCR is obtained when finetuning mDeBERTax on both Italian and Spanish. This is not the case for mBERTx and mDeBERTa, whose performance is hindered by the addition of Italian training data. MCR is much narrower in terms of culinary variety, focusing solely on Colombian recipes. On the other hand, GZ contains not just traditional Italian recipes, but an international range of dishes. This is probably the reason why bilingual training is helpful on GZ, but is not beneficial with relation to MCR: adding data from a separate locale helps the models when approaching the more varied GZ, helping them generalize more effectively over its data. Conversely, they are thrown off by the addition of out-of-domain data when tested on MCR's narrower domain.</p><p>Comparing the EMT fine-tuning results with the baselines at the bottom of Table <ref type="table" target="#tab_6">6</ref>, we can see that further fine-tuning on EMT does provide a boost, compared to training only on XL-WA. Nonetheless, the difference in performance is much greater when testing on GZ, compared to MCR. When looking at mBERTx, fine-tuned on both Italian and Spanish, the model improves by more than 23 Exact points on GZ, while the gap in performance is just under 16 points on MCR. This effect is even more dramatic for mDeBERTax, with a difference of more than 25 points on GZ, but only 2.48 points on MCR.</p><p>Compounded with the fact that, in general, the metrics are much higher when testing on MCR compared to GZ, this points to MCR being a much less challenging test set, compared to GZ. As previously mentioned, part of the dataset is likely machine translated, and since an MT engine is more likely to follow rigidly defined patterns compared to a human translator, this might play a role into the alignment task being easier on these data.</p><p>Table <ref type="table" target="#tab_8">7</ref> reports the performance of the best overall models on each of the individual classes, on both GZ and MCR. Giza++ essentially matches mDeBERTa's performance on MCR, which once again points to entities in MCR being easier to identify compared to GZ. However, the similar performance is largely due to mDeBERTa performing poorly on the unit and part classes, due to the reasons outlined in Section 4. <ref type="table" target="#tab_9">8</ref> reports the results for the NER task for all language settings. For each language, we use the aligner models which obtained the highest results on the entity alignment task. Note that, since the aligner performance does not significantly improve with increased shuffling (see Section 5), we only train aligner models up to 𝑃 = 0.2 for the Spanish setting due to computational constraints.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Entity Recognition Table</head><p>In the Spanish monolingual setting, both BERTes and mBERT obtain F1 scores between 0.92 and 0.95 when fine-tuned on SMT, with the models fine-tuned on EMT trailing behind by 11 to 12 points. As all the models perform similarly and the standard deviation is also close to zero, it once again appears that the entities contained in the MCR dataset are not too challenging for both the mono-and multilingual models to identify. Results of the alignment task by class for the best models, using the Exact metric. Best on GZ in bold, best on MCR underlined.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Model</head><p>In the bilingual fine-tuning scenario, the training data is a concatenation of the SMT datasets produced by the models obtaining the highest performance on the two test sets. Since this is a bilingual fine-tuning scenario, we only use mBERT, as the monolingual models would not be able to be fine-tuned appropriately on this multilingual data. In this setup, the usefulness of the BERT-based aligners becomes more evident. Indeed, while performance on MCR is largely similar to the other setups, with all models outperforming the baseline by a large amount, the same cannot be said for mBERT's performance on GZ. Finetuning mBERT on the combination of the Italian and Spanish data aligned by Fast Align and Giza++ makes the NER model considerably worse at identifying entities in GZ, with a performance decrease of 20 F1 points with the data created by Fast Align and of 21 F1 points with that created by Giza++. The opposite is true when finetuning the mBERT NER model on the SMT data created by mDeBERTa, with the model achieving an F1 of 0.94, beating the baseline by 5 points. Compared to the model fine-tuned on data created by Giza++, this represents a 26 F1 point increase in performance.</p><p>As regards the baseline model fine-tuned on TASTEset's English data and tested on MCR's Spanish entities, we can see that, unexpectedly, the model obtains a 0.88 F1 score, outperforming the mBERT (0.83 F1) and BERTes (0.84 F1) models fine-tuned on the monolingual Spanish EMT data. Despite this, fine-tuning on SMT data produced through our alignment approach allows the NER models to beat this 0.88 F1 baseline, reaching scores as high as 0.95 F1, as previously mentioned.</p><p>In all three scenarios, mBERT achieves performances comparable to those of the monolingual models. This shows that, when inferring on multilingual corpora to extract entities, a single multilingual model can be used, saving time and computational resources both during training and inference.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>B. XL-WA</head><p>As additional data for intermediate word-alignment training, we use XL-WA <ref type="bibr" target="#b15">[16]</ref> </p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Annotation task instructions.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>Dataset class distributions. EMT and SMT refer to the entityand sentence-wise machine-translated TASTEset. GZ refers to our testing dataset.</figDesc><table><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>4,017</cell><cell></cell><cell>5,958</cell><cell>6,473</cell></row><row><cell>qty.</cell><cell></cell><cell cols="2">3,780</cell><cell>3,777</cell><cell></cell><cell>10,186</cell><cell>6,564</cell></row><row><cell>unit</cell><cell></cell><cell cols="2">3,172</cell><cell>3,159</cell><cell></cell><cell>8,148</cell><cell>4,450</cell></row><row><cell cols="2">process</cell><cell cols="2">1,091</cell><cell>1,090</cell><cell></cell><cell>217</cell><cell>265</cell></row><row><cell cols="2">phys. q.</cell><cell></cell><cell>793</cell><cell>791</cell><cell></cell><cell>1,245</cell><cell>1,547</cell></row><row><cell>color</cell><cell></cell><cell></cell><cell>231</cell><cell>231</cell><cell></cell><cell>482</cell><cell>479</cell></row><row><cell>taste</cell><cell></cell><cell></cell><cell>126</cell><cell>125</cell><cell></cell><cell>98</cell><cell>72</cell></row><row><cell cols="2">purpose</cell><cell></cell><cell>94</cell><cell>94</cell><cell></cell><cell>69</cell><cell>126</cell></row><row><cell>part</cell><cell></cell><cell></cell><cell>55</cell><cell>55</cell><cell></cell><cell>220</cell><cell>263</cell></row><row><cell>total</cell><cell></cell><cell cols="2">13,362</cell><cell cols="2">13,259</cell><cell cols="2">26,631 20,272</cell></row><row><cell></cell><cell>s 1</cell><cell></cell><cell>s 2</cell><cell></cell><cell></cell><cell>s 3</cell></row><row><cell cols="2">s 1,1 s 1,2</cell><cell cols="2">s 1,3 s 2,1 s 2,2</cell><cell cols="2">s 2,3 s 3,1 s 3,2</cell><cell>s 3,3</cell></row><row><cell></cell><cell></cell><cell cols="5">M aligner (A, B) = t 1,3 = {'start': 9, 'end': 14}</cell></row><row><cell>t 1,1</cell><cell cols="2">t 1,2 t 1,3 t 2,1</cell><cell>t 2,2</cell><cell>t 2,3</cell><cell>t 3,1</cell><cell>t 3,2</cell><cell>t 3,3</cell></row><row><cell cols="3">t 1 (shuffled)</cell><cell cols="2">t 2 (unshuffled)</cell><cell></cell><cell cols="2">t 3 (shuffled)</cell></row></table><note>A 4 • cups • flour ; 1/2 teaspoon salt ; 1 teaspoon baking soda B farina 4 tazze ; 1/2 cucchiaino sale ; cucchiaino 1 bicarbonato di sodioFigure 1: Aligning source 𝑠 𝑖 and shuffled target 𝑡 𝑗 entities.</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>exemplifies this, where the model 𝑀 𝑎𝑙𝑖𝑔𝑛𝑒𝑟 is trained to predict Exact metric results of the alignment task; averaged out of 5 random runs, besides the XL-WA baseline. Best in bold.</figDesc><table><row><cell>Data</cell><cell>P</cell><cell>mBERT</cell><cell>mBERTx</cell></row><row><cell></cell><cell>0.0</cell><cell>35.93±0.79</cell><cell>38.87±0.48</cell></row><row><cell></cell><cell>0.1</cell><cell>43.13±2.51</cell><cell>44.49±1.21</cell></row><row><cell></cell><cell>0.2</cell><cell>42.54±1.37</cell><cell>44.02±3.32</cell></row><row><cell></cell><cell>0.3</cell><cell>42.49±3.64</cell><cell>46.61±1.62</cell></row><row><cell></cell><cell>0.4</cell><cell>42.31±2.58</cell><cell>47.04±4.01</cell></row><row><cell>EMT</cell><cell>0.5</cell><cell>41.87±1.93</cell><cell>47.22±1.64</cell></row><row><cell></cell><cell>0.6</cell><cell>44.84±2.19</cell><cell>46.89±3.36</cell></row><row><cell></cell><cell>0.7</cell><cell>42.87±3.61</cell><cell>47.36±2.06</cell></row><row><cell></cell><cell>0.8</cell><cell>44.08±1.98</cell><cell>48.34±2.73</cell></row><row><cell></cell><cell>0.9</cell><cell>42.87±3.27</cell><cell>47.28±1.49</cell></row><row><cell></cell><cell>1.0</cell><cell>41.65±2.25</cell><cell>45.98±1.97</cell></row><row><cell>XL-WA</cell><cell>-</cell><cell></cell><cell>21.04</cell></row><row><cell>Data</cell><cell>P</cell><cell>mDeBERTa</cell><cell>mDeBERTax</cell></row><row><cell></cell><cell>0.0</cell><cell>42.17±1.19</cell><cell>46.98±3.77</cell></row><row><cell></cell><cell>0.1</cell><cell>57.00±0.94</cell><cell>58.45±1.37</cell></row><row><cell></cell><cell>0.2</cell><cell>55.03±2.40</cell><cell>57.02±2.43</cell></row><row><cell></cell><cell>0.3</cell><cell>57.09 ± 3.61</cell><cell>60.25±2.35</cell></row><row><cell></cell><cell>0.4</cell><cell>57.26 ± 1.09</cell><cell>59.21±2.59</cell></row><row><cell>EMT</cell><cell>0.5</cell><cell>55.97 ± 3.11</cell><cell>58.43±2.53</cell></row><row><cell></cell><cell cols="2">0.6 58.37 ± 2.46</cell><cell>61.07±2.94</cell></row><row><cell></cell><cell>0.7</cell><cell>57.07 ± 1.58</cell><cell>60.68±3.01</cell></row><row><cell></cell><cell>0.8</cell><cell>57.31 ± 1.20</cell><cell>62.08±3.74</cell></row><row><cell></cell><cell>0.9</cell><cell>56.95 ± 2.69</cell><cell>61.05±1.27</cell></row><row><cell></cell><cell>1.0</cell><cell>57.59 ± 1.81</cell><cell>60.87±1.13</cell></row><row><cell>XL-WA</cell><cell>-</cell><cell></cell><cell>31.71</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Giza++ mBERTx mDeBERTax</head><label></label><figDesc></figDesc><table><row><cell>Qty.</cell><cell>18.41</cell><cell>35.21</cell><cell>30.09</cell><cell>54.95</cell></row><row><cell>Unit</cell><cell>30.94</cell><cell>15.24</cell><cell>24.81</cell><cell>29.75</cell></row><row><cell>Food</cell><cell>61.95</cell><cell>77.01</cell><cell>81.66</cell><cell>83.49</cell></row><row><cell>Process</cell><cell>15.27</cell><cell>51.91</cell><cell>62.60</cell><cell>83.21</cell></row><row><cell>Color</cell><cell>33.70</cell><cell>84.81</cell><cell>67.04</cell><cell>85.93</cell></row><row><cell>Phys. q.</cell><cell>39.00</cell><cell>71.76</cell><cell>61.41</cell><cell>87.66</cell></row><row><cell>Taste</cell><cell>0.00</cell><cell>27.03</cell><cell>35.14</cell><cell>75.68</cell></row><row><cell>Purpose</cell><cell>25.64</cell><cell>61.54</cell><cell>94.87</cell><cell>89.74</cell></row><row><cell>Part</cell><cell>52.48</cell><cell>63.37</cell><cell>13.86</cell><cell>14.85</cell></row><row><cell cols="2">Macro avg. 30.82</cell><cell>54.21</cell><cell>52.38</cell><cell>67.25</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 3</head><label>3</label><figDesc>Exact metric results of the alignment task by class on GZ for the best models (trained on IT⊕ES). Best in bold.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_4"><head>Table 4</head><label>4</label><figDesc>Model performance for the entity recognition task, in terms of F 1 measure. All results are macro avg. out of 5 random runs. number of training instances (55). However, the models obtain high scores on the purpose class, also just 94 instances (mBERTx gets 94.87 Exact score). Unfortunately, repeating the approach we used for the unit class is not feasible, as fine-tuning the model on just 55 instances does not produce any reliable results (𝐸𝑝𝑎𝑟𝑡 = 3.96), meaning this will have to be left for future work.The rest of the results from Table3are generally in line with the average results from Table2. The scores achieved by the baselines for each class do not have any evident outliers, save for Fast Align scoring a 0 on taste.</figDesc><table><row><cell>Train</cell><cell>Test</cell><cell>Aligner</cell><cell>NER</cell><cell>F 1</cell></row><row><cell></cell><cell></cell><cell>-</cell><cell></cell><cell>0.89±0.01</cell></row><row><cell></cell><cell></cell><cell>mBERTx</cell><cell></cell><cell>0.91±0.02</cell></row><row><cell></cell><cell></cell><cell>mDeBERTax</cell><cell>mBERT</cell><cell>0.94±0.01</cell></row><row><cell></cell><cell></cell><cell>Fast Align</cell><cell></cell><cell>0.84±0.01</cell></row><row><cell>it</cell><cell>it</cell><cell>Giza++ -</cell><cell></cell><cell>0.87±0.03 0.86±0.01</cell></row><row><cell></cell><cell></cell><cell>mBERTx</cell><cell></cell><cell>0.9±0.04</cell></row><row><cell></cell><cell></cell><cell>mDeBERTax Fast Align</cell><cell>BERT it</cell><cell>0.94±0.0 0.85±0.04</cell></row><row><cell></cell><cell></cell><cell>Giza++</cell><cell></cell><cell>0.91±0.03</cell></row><row><cell>en</cell><cell>it en</cell><cell>-</cell><cell>mBERT</cell><cell>0.79±0.05 0.9±0.01</cell></row><row><cell></cell><cell>en</cell><cell></cell><cell>BERTen</cell><cell>0.91±0.01</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_5"><head>Table 5</head><label>5</label><figDesc>Dataset class distributions. EMT and SMT refer to the entity-and sentence-wise machine-translated TASTEset. GZ and MCR refer to our testing datasets.</figDesc><table><row><cell></cell><cell></cell><cell cols="2">TS / EMT</cell><cell></cell><cell></cell><cell cols="2">SMT</cell><cell></cell><cell>GZ</cell><cell>MCR</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell cols="2">mBERTx</cell><cell></cell><cell cols="2">mDeBERTax</cell><cell></cell></row><row><cell cols="2">Class</cell><cell cols="2">en / it / es</cell><cell>it</cell><cell>es</cell><cell></cell><cell>it</cell><cell>es</cell><cell>en</cell><cell>it</cell><cell>en</cell><cell>es</cell></row><row><cell cols="2">food</cell><cell></cell><cell>4,020</cell><cell>3,999</cell><cell cols="2">4,012</cell><cell>4,017</cell><cell>4,018</cell><cell>5,958</cell><cell>6,473</cell><cell>3,600</cell><cell>1,143</cell></row><row><cell cols="2">quantity</cell><cell></cell><cell>3,780</cell><cell>3,764</cell><cell cols="2">3,778</cell><cell>3,777</cell><cell>3,780</cell><cell>10,186</cell><cell>6,564</cell><cell>2,945</cell><cell>962</cell></row><row><cell cols="2">unit</cell><cell></cell><cell>3,172</cell><cell>3,151</cell><cell cols="2">3,169</cell><cell>3,159</cell><cell>3,171</cell><cell>8,148</cell><cell>4,450</cell><cell>2,325</cell><cell>760</cell></row><row><cell cols="2">process</cell><cell></cell><cell>1,091</cell><cell>1,066</cell><cell cols="2">1,089</cell><cell>1,090</cell><cell>1,091</cell><cell>217</cell><cell>265</cell><cell>1,236</cell><cell>379</cell></row><row><cell cols="3">physical q.</cell><cell>793</cell><cell>785</cell><cell cols="2">791</cell><cell>791</cell><cell>793</cell><cell>1,245</cell><cell>1,547</cell><cell>897</cell><cell>285</cell></row><row><cell cols="2">color</cell><cell></cell><cell>231</cell><cell>226</cell><cell cols="2">231</cell><cell>231</cell><cell>231</cell><cell>482</cell><cell>479</cell><cell>309</cell><cell>97</cell></row><row><cell cols="2">taste</cell><cell></cell><cell>126</cell><cell>121</cell><cell cols="2">123</cell><cell>125</cell><cell>123</cell><cell>98</cell><cell>72</cell><cell>8</cell><cell>2</cell></row><row><cell cols="2">purpose</cell><cell></cell><cell>94</cell><cell>94</cell><cell></cell><cell>94</cell><cell>94</cell><cell>94</cell><cell>69</cell><cell>126</cell><cell>89</cell><cell>34</cell></row><row><cell cols="2">part</cell><cell></cell><cell>55</cell><cell>53</cell><cell></cell><cell>55</cell><cell>55</cell><cell>55</cell><cell>220</cell><cell>263</cell><cell>142</cell><cell>44</cell></row><row><cell cols="2">total</cell><cell cols="2">13,362</cell><cell cols="3">13,259 13,342</cell><cell cols="2">13,339 13,356</cell><cell cols="2">26,631 20,272</cell><cell>11,551 3,706</cell></row><row><cell>Data</cell><cell>P</cell><cell>GZ</cell><cell cols="2">mBERT MCR</cell><cell>GZ</cell><cell cols="2">mBERTx MCR</cell><cell cols="3">mDeBERTa GZ MCR</cell><cell>mDeBERTax GZ MCR</cell></row><row><cell>EMT</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell></cell><cell>0</cell><cell cols="2">35.93±0.79</cell><cell></cell><cell cols="2">38.87±0.48</cell><cell></cell><cell cols="2">42.17±1.19</cell><cell>46.98±3.77</cell></row><row><cell>it</cell><cell cols="3">0.1 43.13±2.51</cell><cell></cell><cell cols="2">44.49±1.21</cell><cell></cell><cell cols="2">57.00±0.94</cell><cell>58.45±1.37</cell></row><row><cell></cell><cell cols="3">0.2 42.54±1.37</cell><cell></cell><cell cols="2">44.02±3.32</cell><cell></cell><cell cols="2">55.03±2.40</cell><cell>57.02±2.43</cell></row><row><cell></cell><cell>0</cell><cell></cell><cell cols="2">49.03±0.59</cell><cell></cell><cell></cell><cell>50.38±1.10</cell><cell></cell><cell cols="2">51.93±0.65</cell><cell>53.20±0.83</cell></row><row><cell>es</cell><cell>0.1</cell><cell></cell><cell cols="2">63.69±0.96</cell><cell></cell><cell></cell><cell>67.60±0.74</cell><cell></cell><cell cols="2">70.43±2.48</cell><cell>71.07±3.62</cell></row><row><cell></cell><cell>0.2</cell><cell></cell><cell cols="2">66.07±1.30</cell><cell></cell><cell></cell><cell>70.20±1.66</cell><cell></cell><cell cols="2">69.25±1.94</cell><cell>72.62±1.93</cell></row><row><cell></cell><cell>0</cell><cell cols="9">33.82±5.30 46.59±0.98 41.54±1.87 47.72±1.56 46.98±3.77 40.33±1.97 45.17±2.58 52.70±1.09</cell></row><row><cell>it-es</cell><cell cols="10">0.1 43.36±2.72 64.57±2.35 46.14±3.85 67.16±2.19 58.45±1.37 53.68±2.45 57.64±0.83 72.95±1.75</cell></row><row><cell></cell><cell cols="10">0.2 44.37±1.57 67.62±0.33 47.14±1.43 69.10±1.10 57.02±2.43 54.87±1.37 58.84±2.68 72.71±1.83</cell></row><row><cell>XL-WA</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell></row><row><cell>it</cell><cell>-</cell><cell></cell><cell></cell><cell></cell><cell cols="2">21.04</cell><cell></cell><cell></cell><cell></cell><cell>31.71</cell></row><row><cell>es</cell><cell>-</cell><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>54.14</cell><cell></cell><cell></cell><cell>58.56</cell></row><row><cell>it-es</cell><cell>-</cell><cell></cell><cell></cell><cell></cell><cell cols="2">23.60</cell><cell>53.89</cell><cell></cell><cell></cell><cell>33.56 70.47</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_6"><head>Table 6</head><label>6</label><figDesc>Alignment task results (Exact metric). All results are averaged out of 5 random runs, besides the XL-WA baselines. Best in bold.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_7"><head>Test set Qty. Unit Food Process Color Phys. q. Taste Purpose Part Macro avg.</head><label></label><figDesc></figDesc><table><row><cell>Fast Align</cell><cell>GZ</cell><cell>18.41 30.94 61.95 15.27</cell><cell>33.70</cell><cell>39.00</cell><cell>0.00</cell><cell>25.64</cell><cell>52.48</cell><cell>30.82</cell></row><row><cell>Fast Align</cell><cell cols="2">MCR 54.27 71.82 62.73 42.77</cell><cell>66.67</cell><cell>45.68</cell><cell>0.00</cell><cell>58.82</cell><cell>40.00</cell><cell>49.20</cell></row><row><cell>Giza++</cell><cell>GZ</cell><cell>35.21 15.24 77.01 51.91</cell><cell>84.81</cell><cell>71.76</cell><cell>27.03</cell><cell>61.54</cell><cell>63.37</cell><cell>54.21</cell></row><row><cell>Giza++</cell><cell cols="2">MCR 90.29 89.31 76.93 76.30</cell><cell>79.76</cell><cell>75.72</cell><cell>50.00</cell><cell>82.35</cell><cell>68.57</cell><cell>76.58</cell></row><row><cell cols="7">GZ MCR 95.30 3.93 89.32 81.72 30.09 24.81 81.66 62.60 54.95 29.75 83.49 83.21 85.93 87.66 75.68 67.04 61.41 35.14 87.50 77.02 100.00 100.00 94.87 89.74 mDeBERTax en-it-es GZ mBERTx en-it-es mBERTx en-it-es mDeBERTa MCR 97.05 11.25 90.48 93.91 94.32 93.95 100.00 97.06</cell><cell>13.86 9.52 14.85 14.29</cell><cell>52.38 71.59 67.25 76.92</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_8"><head>Table 7</head><label>7</label><figDesc></figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_9"><head>Table 8</head><label>8</label><figDesc>, a multilingual word-alignment Entity recognition task F 1 scores (5 random runs macro avg).dataset<ref type="bibr" target="#b15">[16]</ref> built from WikiMatrix<ref type="bibr" target="#b48">[48]</ref>,12 featuring 14 EN-XX language combinations. Its training set is composed of silver labels generated by a statistical model, while the development and test sets are manually annotated. Since XL-WA has a balanced domain distribution and can be considered representative of general language, it can be a good resource on which to train a baseline word-alignment model. Table9reports statistics for the EN-IT and EN-ES partitions used in this study.</figDesc><table><row><cell>Train</cell><cell>Test</cell><cell>Aligner</cell><cell>NER</cell><cell>F 1</cell></row><row><cell></cell><cell></cell><cell>-</cell><cell></cell><cell>0.89±0.01</cell></row><row><cell></cell><cell></cell><cell>mBERT</cell><cell></cell><cell>0.91±0.02</cell></row><row><cell></cell><cell></cell><cell>mDeBERTa</cell><cell>mBERT</cell><cell>0.94±0.01</cell></row><row><cell></cell><cell></cell><cell>Fast Align</cell><cell></cell><cell>0.84±0.01</cell></row><row><cell>it</cell><cell>it</cell><cell>Giza++ -</cell><cell></cell><cell>0.87±0.03 0.86±0.01</cell></row><row><cell></cell><cell></cell><cell>mBERT</cell><cell></cell><cell>0.9±0.04</cell></row><row><cell></cell><cell></cell><cell>mDeBERTa Fast Align</cell><cell>BERT it</cell><cell>0.94±0.0 0.85±0.04</cell></row><row><cell></cell><cell></cell><cell>Giza++</cell><cell></cell><cell>0.91±0.03</cell></row><row><cell></cell><cell></cell><cell>-</cell><cell></cell><cell>0.83±0.01</cell></row><row><cell></cell><cell></cell><cell>mBERT</cell><cell></cell><cell>0.95±0.0</cell></row><row><cell></cell><cell></cell><cell>mDeBERTa</cell><cell>mBERT</cell><cell>0.92±0.01</cell></row><row><cell></cell><cell></cell><cell>Fast Align</cell><cell></cell><cell>0.94±0.0</cell></row><row><cell>es</cell><cell>es</cell><cell>Giza++ -</cell><cell></cell><cell>0.95±0.0 0.84±0.0</cell></row><row><cell></cell><cell></cell><cell>mBERT</cell><cell></cell><cell>0.95±0.0</cell></row><row><cell></cell><cell></cell><cell>mDeBERTa</cell><cell>BERTes</cell><cell>0.93±0.01</cell></row><row><cell></cell><cell></cell><cell>Fast Align</cell><cell></cell><cell>0.95±0.0</cell></row><row><cell></cell><cell></cell><cell>Giza++</cell><cell></cell><cell>0.95±0.0</cell></row><row><cell></cell><cell></cell><cell>-</cell><cell></cell><cell>0.89±0.01</cell></row><row><cell></cell><cell>it</cell><cell>Fast Align Giza++</cell><cell>mBERT</cell><cell>0.69±0.01 0.68±0.03</cell></row><row><cell>it-es</cell><cell></cell><cell>mDeBERTa -</cell><cell></cell><cell>0.94±0.01 0.83±0.0</cell></row><row><cell></cell><cell>es</cell><cell>Fast Align Giza++</cell><cell>mBERT</cell><cell>0.95±0.0 0.95±0.0</cell></row><row><cell></cell><cell></cell><cell>mDeBERTa</cell><cell></cell><cell>0.94±0.01</cell></row><row><cell></cell><cell>it</cell><cell></cell><cell></cell><cell>0.79±0.05</cell></row><row><cell>en</cell><cell>es en (GZ) en (MCR)</cell><cell>-</cell><cell>mBERT</cell><cell>0.88±0.01 0.9±0.01 0.93±0.0</cell></row><row><cell></cell><cell>en (GZ) en (MCR)</cell><cell></cell><cell>BERTen</cell><cell>0.91±0.01 0.93±0.0</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_10"><head>Table 9</head><label>9</label><figDesc>Statistics for XL-WA's EN-IT and EN-ES subsets. All models are trained on a single NVIDIA RTX 5000 Ada Generation, with 32 GB of VRAM. The total training time is around 7-15 minutes for each alignment model, depending on the training data combination, plus 30-60 minutes for training each on XL-WA. Training each NER model takes around 6-7 minutes. All the training, including multiple models for standard deviation calculation, was carried out in under 48 hours.</figDesc><table /><note>C. Computational Resources</note></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_0">Resources available at https://github.com/paolo-gajo/food</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_1">https://www.allrecipes.com</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_2">https://www.deepl.com/en/docs-api</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_3">The effect on model performance upon training is negligible given that these predictions constitute less than 1% of the total.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_4">https://www.giallozafferano.it</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_5">https://labelstud.io</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_6">We do not use the larger mDeBERTa model due to the computational cost deriving from the number of language combinations.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_7"><ref type="bibr" target="#b8">9</ref> https://huggingface.co/docs/transformers/en/tasks/token_ classification</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_8">https://huggingface.co/docs/transformers/en/main_classes/ tokenizer#transformers.BatchEncoding.char_to_token</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="11" xml:id="foot_9">https://www.mycolombianrecipes.com</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="12" xml:id="foot_10">https://ai.meta.com/blog/wikimatrix/</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">SemEval-2022 task 11: Multilingual complex named entity recognition (MultiCoNER)</title>
		<author>
			<persName><forename type="first">S</forename><surname>Malmasi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Fang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Fetahu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Rokhlenko</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/2022.semeval-1.196" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 16th International Workshop on Semantic Evaluation (SemEval-2022)</title>
				<editor>
			<persName><forename type="first">G</forename><surname>Emerson</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Schluter</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Stanovsky</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Kumar</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Palmer</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Schneider</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Singh</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Ratan</surname></persName>
		</editor>
		<meeting>the 16th International Workshop on Semantic Evaluation (SemEval-2022)<address><addrLine>Seattle, United States</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="1412" to="1437" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">SemEval-2023 task 2: Fine-grained multilingual named entity recognition (MultiCoNER 2)</title>
		<author>
			<persName><forename type="first">B</forename><surname>Fetahu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Rokhlenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Malmasi</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/2023.semeval-1.310" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 17th International Workshop on Semantic Evaluation (SemEval-2023), Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Ojha</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Doğruöz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">G</forename><surname>Da San Martino</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Tayyar</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Madabushi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><surname>Kumar</surname></persName>
		</editor>
		<editor>
			<persName><surname>Sartori</surname></persName>
		</editor>
		<meeting>the 17th International Workshop on Semantic Evaluation (SemEval-2023), Association for Computational Linguistics<address><addrLine>Toronto, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="2247" to="2265" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Mise en Place: Unsupervised Interpretation of Instructional Recipes</title>
		<author>
			<persName><forename type="first">C</forename><surname>Kiddon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">T</forename><surname>Ponnuraj</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zettlemoyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Choi</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/D15-1114</idno>
		<ptr target="https://aclanthology.org/D15-1114.doi:10.18653/v1/D15-1114" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">L</forename><surname>Màrquez</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Callison-Burch</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Su</surname></persName>
		</editor>
		<meeting>the 2015 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics<address><addrLine>Lisbon, Portugal</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="982" to="992" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">English Recipe Flow Graph Corpus</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Yamakata</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mori</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Carroll</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/2020.lrec-1.638" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Twelfth Language Resources and Evaluation Conference, European Language Resources Association</title>
				<editor>
			<persName><forename type="first">N</forename><surname>Calzolari</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Béchet</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">P</forename><surname>Blache</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">K</forename><surname>Choukri</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Cieri</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Declerck</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Goggi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Isahara</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Maegaard</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Mariani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Mazo</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Moreno</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Odijk</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Piperidis</surname></persName>
		</editor>
		<meeting>the Twelfth Language Resources and Evaluation Conference, European Language Resources Association<address><addrLine>Marseille, France</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="5187" to="5194" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Learning Program Representations for Food Images and Cooking Recipes</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">P</forename><surname>Papadopoulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Mora</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Chepurko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">W</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Ofli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Torralba</surname></persName>
		</author>
		<idno type="DOI">10.1109/CVPR52688.2022.01606</idno>
		<ptr target="https://ieeexplore.ieee.org/document/9878478/.doi:10.1109/CVPR52688.2022.01606" />
	</analytic>
	<monogr>
		<title level="m">IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)</title>
				<meeting><address><addrLine>New Orleans, LA, USA</addrLine></address></meeting>
		<imprint>
			<publisher>IEEE</publisher>
			<date type="published" when="2022">2022. 2022</date>
			<biblScope unit="page" from="16538" to="16548" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">End-to-end Parsing of Procedural Text into Flow Graphs</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">J</forename><surname>Bhatt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">A</forename><surname>Hosseini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Fancellu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Fazly</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/2024.lrec-main.517" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)</title>
				<editor>
			<persName><forename type="first">N</forename><surname>Calzolari</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M.-Y</forename><surname>Kan</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">V</forename><surname>Hoste</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Lenci</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Sakti</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Xue</surname></persName>
		</editor>
		<meeting>the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)<address><addrLine>Torino, Italia</addrLine></address></meeting>
		<imprint>
			<publisher>ELRA and ICCL</publisher>
			<date type="published" when="2024">2024</date>
			<biblScope unit="page" from="5833" to="5842" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">Unsupervised Learning of Graph from Recipes</title>
		<author>
			<persName><forename type="first">A</forename><surname>Diallo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bikakis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Dickens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Hunter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Miller</surname></persName>
		</author>
		<ptr target="http://arxiv.org/abs/2401.12088" />
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Entity projection via machine translation for cross-lingual NER</title>
		<author>
			<persName><forename type="first">A</forename><surname>Jain</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Paranjape</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><forename type="middle">C</forename><surname>Lipton</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/D19-1100</idno>
		<ptr target="https://aclanthology.org/D19-1100.doi:10.18653/v1/D19-1100" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">K</forename><surname>Inui</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Jiang</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">V</forename><surname>Ng</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">X</forename><surname>Wan</surname></persName>
		</editor>
		<meeting>the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), Association for Computational Linguistics<address><addrLine>Hong Kong, China</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="1083" to="1092" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">A systematic comparison of various statistical alignment models</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">J</forename><surname>Och</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Ney</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Computational Linguistics</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="page" from="19" to="51" />
			<date type="published" when="2003">2003</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">A Simple, Fast, and Effective Reparameterization of IBM Model 2</title>
		<author>
			<persName><forename type="first">C</forename><surname>Dyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Chahuneau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">A</forename><surname>Smith</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/N13-1073" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">L</forename><surname>Vanderwende</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Daumé</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Iii</forename></persName>
		</editor>
		<editor>
			<persName><forename type="first">K</forename><surname>Kirchhoff</surname></persName>
		</editor>
		<meeting>the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Association for Computational Linguistics<address><addrLine>Atlanta, Georgia</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="644" to="648" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">The mathematics of statistical machine translation: Parameter estimation</title>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">F</forename><surname>Brown</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">A</forename><surname>Della Pietra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">J</forename><surname>Della Pietra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">L</forename><surname>Mercer</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/J93-2003" />
	</analytic>
	<monogr>
		<title level="j">Computational Linguistics</title>
		<imprint>
			<biblScope unit="volume">19</biblScope>
			<biblScope unit="page" from="263" to="311" />
			<date type="published" when="1993">1993</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Attention is All you Need</title>
		<author>
			<persName><forename type="first">A</forename><surname>Vaswani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shazeer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Parmar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Uszkoreit</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Jones</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">N</forename><surname>Gomez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Kaiser</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Polosukhin</surname></persName>
		</author>
		<ptr target="https://proceedings.neurips.cc/paper_files/paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html" />
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems</title>
				<imprint>
			<publisher>Curran Associates, Inc</publisher>
			<date type="published" when="2017">2017</date>
			<biblScope unit="volume">30</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">BERT: Pre-training of deep bidirectional transformers for language understanding</title>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-W</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/N19-1423</idno>
		<ptr target="https://aclanthology.org/N19-1423.doi:10.18653/v1/N19-1423" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</title>
		<title level="s">Long and Short Papers</title>
		<editor>
			<persName><forename type="first">J</forename><surname>Burstein</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><surname>Doran</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Solorio</surname></persName>
		</editor>
		<meeting>the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies<address><addrLine>Minneapolis, Minnesota</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="4171" to="4186" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">A supervised word alignment method based on cross-language span prediction using multilingual BERT</title>
		<author>
			<persName><forename type="first">M</forename><surname>Nagata</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Chousa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Nishino</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2020.emnlp-main.41</idno>
		<ptr target="https://aclanthology.org/2020.emnlp-main.41.doi:10.18653/v1/2020.emnlp-main.41" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">B</forename><surname>Webber</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Cohn</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>He</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</editor>
		<meeting>the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="555" to="565" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<title level="m" type="main">Cross-Lingual Named Entity Recognition Using Parallel Corpus: A New Approach Using XLM-RoBERTa Alignment</title>
		<author>
			<persName><forename type="first">B</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Xu</surname></persName>
		</author>
		<ptr target="http://arxiv.org/abs/2101.11112" />
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">XL-WA: a Gold Evaluation Benchmark for Word Alignment in 14 Language Pairs</title>
		<author>
			<persName><forename type="first">F</forename><surname>Martelli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Bejgu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Campagnano</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Čibej</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Costa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gantar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kallas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Koeva</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Koppel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Krek</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Langemets</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Lipp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Nimb</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Olsen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">S</forename><surname>Pedersen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Quochi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Salgado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Simon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Tiberius</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R.-J</forename><surname>Ureña-Ruiz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Navigli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Ninth Italian Conference on Computational Linguistics (CLiC-it 2023)</title>
				<editor>
			<persName><forename type="first">F</forename><surname>Boschetti</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><forename type="middle">N</forename><surname>Gianluca</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">E</forename><surname>Lebani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Bernardo</forename><surname>Magnini</surname></persName>
		</editor>
		<meeting>the Ninth Italian Conference on Computational Linguistics (CLiC-it 2023)<address><addrLine>Venice, Italy</addrLine></address></meeting>
		<imprint>
			<publisher>CEUR-WS</publisher>
			<date type="published" when="2023">2023</date>
			<biblScope unit="volume">3596</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Cross-lingual semantic role labeling with high-quality translated training corpus</title>
		<author>
			<persName><forename type="first">H</forename><surname>Fei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Ji</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2020.acl-main.627</idno>
		<ptr target="https://aclanthology.org/2020.acl-main.627.doi:10.18653/v1/2020.acl-main.627" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">D</forename><surname>Jurafsky</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Chai</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Schluter</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Tetreault</surname></persName>
		</editor>
		<meeting>the 58th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="7014" to="7026" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Model and data transfer for cross-lingual sequence labelling in zero-resource settings</title>
		<author>
			<persName><forename type="first">I</forename><surname>García-Ferrero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Agerri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Rigau</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/2022.findings-emnlp" />
	</analytic>
	<monogr>
		<title level="m">Findings of the Association for Computational Linguistics: EMNLP 2022, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">Y</forename><surname>Goldberg</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Z</forename><surname>Kozareva</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</editor>
		<meeting><address><addrLine>Abu Dhabi, United Arab Emirates</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2022">2022</date>
			<biblScope unit="page" from="6403" to="6416" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title/>
		<idno type="DOI">10.18653/v1/2022.findings-emnlp.478</idno>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Word alignment by finetuning embeddings on parallel corpora</title>
		<author>
			<persName><forename type="first">Z.-Y</forename><surname>Dou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Neubig</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2021.eacl-main.181</idno>
		<ptr target="https://aclanthology.org/2021.eacl-main.181.doi:10.18653/v1/2021.eacl-main.181" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume</title>
		<title level="s">Association for Computational Linguistics</title>
		<editor>
			<persName><forename type="first">P</forename><surname>Merlo</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Tiedemann</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Tsarfaty</surname></persName>
		</editor>
		<meeting>the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="2112" to="2128" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Introduction to the CoNLL-2003 Shared Task: Language-Independent Named Entity Recognition</title>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">F</forename><surname>Tjong Kim Sang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">De</forename><surname>Meulder</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/W03-0419" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Seventh Conference on Natural Language Learning at HLT-NAACL 2003</title>
				<meeting>the Seventh Conference on Natural Language Learning at HLT-NAACL 2003</meeting>
		<imprint>
			<date type="published" when="2003">2003</date>
			<biblScope unit="page" from="142" to="147" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Parallel data, tools and interfaces in OPUS</title>
		<author>
			<persName><forename type="first">J</forename><surname>Tiedemann</surname></persName>
		</author>
		<ptr target="http://www.lrec-conf.org/proceedings/lrec2012/pdf/463_Paper.pdf" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Eighth International Conference on Language Resources and Evaluation (LREC&apos;12), European Language Resources Association (ELRA)</title>
				<editor>
			<persName><forename type="first">N</forename><surname>Calzolari</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">K</forename><surname>Choukri</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">T</forename><surname>Declerck</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><forename type="middle">U</forename><surname>Doğan</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">B</forename><surname>Maegaard</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Mariani</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Moreno</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Odijk</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Piperidis</surname></persName>
		</editor>
		<meeting>the Eighth International Conference on Language Resources and Evaluation (LREC&apos;12), European Language Resources Association (ELRA)<address><addrLine>Istanbul, Turkey</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2012">2012</date>
			<biblScope unit="page" from="2214" to="2218" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<title level="m" type="main">Language Models are Few-Shot Learners</title>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">B</forename><surname>Brown</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Mann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Ryder</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Subbiah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kaplan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Dhariwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Neelakantan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Shyam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Sastry</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Askell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Herbert-Voss</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Krueger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Henighan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Child</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ramesh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">M</forename><surname>Ziegler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Winter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Hesse</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Sigler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Litwin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gray</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chess</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Clark</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Berner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mccandlish</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Radford</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sutskever</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Amodei</surname></persName>
		</author>
		<idno>arXiv:</idno>
		<ptr target="2005.14165" />
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<monogr>
		<title level="m" type="main">PaLM: Scaling Language Modeling with Pathways</title>
		<author>
			<persName><forename type="first">A</forename><surname>Chowdhery</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Narang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Devlin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bosma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Mishra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Roberts</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Barham</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">W</forename><surname>Chung</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Sutton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Gehrmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Schuh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Shi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Tsvyashchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Maynez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Barnes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Tay</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Shazeer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Prabhakaran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Reif</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Du</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Hutchinson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Pope</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Bradbury</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Austin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Isard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Gur-Ari</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Yin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Duke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Levskaya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ghemawat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Dev</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Michalewski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Garcia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Misra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Robinson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Fedus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Ippolito</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Luan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Lim</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Zoph</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Spiridonov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Sepassi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dohan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Agrawal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Omernick</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">M</forename><surname>Dai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">S</forename><surname>Pillai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Pellat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lewkowycz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Moreira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Child</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Polozov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Lee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Saeta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Diaz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Firat</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Catasta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Meier-Hellstern</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Eck</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Dean</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Petrov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Fiedel</surname></persName>
		</author>
		<idno>arXiv:</idno>
		<ptr target="2204.02311[cs" />
		<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<title level="m" type="main">LLaMA: Open and Efficient Foundation Language Models</title>
		<author>
			<persName><forename type="first">H</forename><surname>Touvron</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lavril</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Izacard</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Martinet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M.-A</forename><surname>Lachaux</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Lacroix</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Rozière</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Hambro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Azhar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rodriguez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Joulin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Grave</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Lample</surname></persName>
		</author>
		<idno>arXiv:</idno>
		<ptr target="2302.13971" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<monogr>
		<title level="m" type="main">PromptNER: Prompting For Named Entity Recognition</title>
		<author>
			<persName><forename type="first">D</forename><surname>Ashok</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><forename type="middle">C</forename><surname>Lipton</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2305.15444</idno>
		<ptr target="http://arxiv.org/abs/2305.15444" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Chainof-Thought Prompting Elicits Reasoning in Large Language Models</title>
		<author>
			<persName><forename type="first">J</forename><surname>Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Schuurmans</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Bosma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Ichter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Xia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Chi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Le</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Zhou</surname></persName>
		</author>
		<idno>arXiv:</idno>
		<ptr target="2201.11903[cs" />
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<monogr>
		<title level="m" type="main">GPT-NER: Named Entity Recognition via Large Language Models</title>
		<author>
			<persName><forename type="first">S</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Ouyang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Wang</surname></persName>
		</author>
		<idno>arXiv:</idno>
		<ptr target="2304.10428" />
		<imprint>
			<date type="published" when="2023">2023</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<monogr>
		<title level="m" type="main">A Survey on In-context Learning</title>
		<author>
			<persName><forename type="first">Q</forename><surname>Dong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Dai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Zheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Xia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Xu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Sui</surname></persName>
		</author>
		<idno>arXiv:</idno>
		<ptr target="2301.00234[cs" />
		<imprint>
			<date type="published" when="2024">2024</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<analytic>
		<title level="a" type="main">Towards Robust Linguistic Analysis using OntoNotes</title>
		<author>
			<persName><forename type="first">S</forename><surname>Pradhan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Moschitti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Xue</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">T</forename><surname>Ng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Björkelund</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Uryupina</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhong</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/W13-3516" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Seventeenth Conference on Computational Natural Language Learning, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">J</forename><surname>Hockenmaier</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Riedel</surname></persName>
		</editor>
		<meeting>the Seventeenth Conference on Computational Natural Language Learning, Association for Computational Linguistics<address><addrLine>Sofia, Bulgaria</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="143" to="152" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<analytic>
		<title level="a" type="main">Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks</title>
		<author>
			<persName><forename type="first">P</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Perez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Piktus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Petroni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Karpukhin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Goyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Küttler</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>-T. Yih</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Rocktäschel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Riedel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Kiela</surname></persName>
		</author>
		<ptr target="https://proceedings.neurips.cc/paper_files/paper/2020/file/6b493230205f780e1bc26945df7481e5-Paper.pdf" />
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems</title>
				<editor>
			<persName><forename type="first">H</forename><surname>Larochelle</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Ranzato</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Hadsell</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><forename type="middle">F</forename><surname>Balcan</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">H</forename><surname>Lin</surname></persName>
		</editor>
		<imprint>
			<publisher>Curran Associates, Inc</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">33</biblScope>
			<biblScope unit="page" from="9459" to="9474" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">GNN-SL: Sequence Labeling Based on Nearest Examples via GNN</title>
		<author>
			<persName><forename type="first">S</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Meng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Ouyang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Li</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Lyu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Wang</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2023.findings-acl.803</idno>
		<ptr target="https://aclanthology.org/2023.findings-acl.803.doi:10.18653/v1/2023.findings-acl.803" />
	</analytic>
	<monogr>
		<title level="m">Findings of the Association for Computational Linguistics: ACL 2023, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">A</forename><surname>Rogers</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Boyd-Graber</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Okazaki</surname></persName>
		</editor>
		<meeting><address><addrLine>Toronto, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2023">2023</date>
			<biblScope unit="page" from="12679" to="12692" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b32">
	<analytic>
		<title level="a" type="main">RecipeDB: a resource for exploring recipes</title>
		<author>
			<persName><forename type="first">D</forename><surname>Batra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Diwan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Upadhyay</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">S</forename><surname>Kalra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Sharma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">K</forename><surname>Sharma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Khanna</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">S</forename><surname>Marwah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Kalathil</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Singh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Tuwani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Bagler</surname></persName>
		</author>
		<idno type="DOI">10.1093/database/baaa077</idno>
		<ptr target="https://doi.org/10.1093/database/baaa077.doi:10.1093/database/baaa077" />
	</analytic>
	<monogr>
		<title level="j">Database</title>
		<imprint>
			<biblScope unit="page">77</biblScope>
			<date type="published" when="2020">2020. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b33">
	<analytic>
		<title level="a" type="main">FoodOn: a harmonized food ontology to increase global food traceability, quality control and data integration</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">M</forename><surname>Dooley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><forename type="middle">J</forename><surname>Griffiths</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">S</forename><surname>Gosal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">L</forename><surname>Buttigieg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Hoehndorf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">C</forename><surname>Lange</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">M</forename><surname>Schriml</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">S L</forename><surname>Brinkman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">W L</forename><surname>Hsiao</surname></persName>
		</author>
		<idno type="DOI">10.1038/s41538-018-0032-6</idno>
		<idno>doi:</idno>
		<ptr target="10.1038/s41538-018-0032-6" />
	</analytic>
	<monogr>
		<title level="j">npj Science of Food</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page">23</biblScope>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b34">
	<analytic>
		<title level="a" type="main">FoodKG: A Semantics-Driven Knowledge Graph for Food Recommendation</title>
		<author>
			<persName><forename type="first">S</forename><surname>Haussmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Seneviratne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Ne'eman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Codella</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-H</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">L</forename><surname>Mcguinness</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Zaki</surname></persName>
		</author>
		<idno type="DOI">10.1007/978-3-030-30796-7_10</idno>
	</analytic>
	<monogr>
		<title level="m">The Semantic Web -ISWC 2019</title>
				<editor>
			<persName><forename type="first">C</forename><surname>Ghidini</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">O</forename><surname>Hartig</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Maleshkova</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">V</forename><surname>Svátek</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">I</forename><surname>Cruz</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Hogan</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Song</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Lefrançois</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><surname>Gandon</surname></persName>
		</editor>
		<meeting><address><addrLine>Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer International Publishing</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="146" to="162" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b35">
	<monogr>
		<title level="m" type="main">Recipe1M+: A Dataset for Learning Cross-Modal Embeddings for Cooking Recipes and Food Images</title>
		<author>
			<persName><forename type="first">J</forename><surname>Marin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Biswas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Ofli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Hynes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Salvador</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Aytar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Weber</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Torralba</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.1810.06553</idno>
		<idno type="arXiv">arXiv:1810.06553</idno>
		<ptr target="http://arxiv.org/abs/1810.06553.doi:10.48550/arXiv.1810.06553" />
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b36">
	<analytic>
		<title level="a" type="main">RecipeNLG: A Cooking Recipes Dataset for Semi-Structured Text Generation</title>
		<author>
			<persName><forename type="first">M</forename><surname>Bień</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Gilski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Maciejewska</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Taisner</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Wisniewski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Lawrynowicz</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2020.inlg-1.4</idno>
		<ptr target="https://aclanthology.org/2020.inlg-1.4.doi:10.18653/v1/2020.inlg-1.4" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 13th International Conference on Natural Language Generation, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">B</forename><surname>Davis</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>Graham</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Kelleher</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>Sripada</surname></persName>
		</editor>
		<meeting>the 13th International Conference on Natural Language Generation, Association for Computational Linguistics<address><addrLine>Dublin, Ireland</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="22" to="28" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b37">
	<analytic>
		<title level="a" type="main">SMPT: A Semi-Supervised Multi-Model Prediction Technique for Food Ingredient Named Entity Recognition (FINER) Dataset Construction</title>
		<author>
			<persName><forename type="first">K</forename><forename type="middle">S</forename><surname>Komariah</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">T</forename><surname>Purnomo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Satriawan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">O</forename><surname>Hasanuddin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Setianingsih</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B.-K</forename><surname>Sin</surname></persName>
		</author>
		<idno type="DOI">10.3390/informatics10010010</idno>
		<ptr target="https://www.mdpi.com/2227-9709/10/1/10.doi:10.3390/informatics10010010" />
	</analytic>
	<monogr>
		<title level="j">Informatics</title>
		<imprint>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page">10</biblScope>
			<date type="published" when="2023">2023</date>
		</imprint>
		<respStmt>
			<orgName>Multidisciplinary Digital Publishing Institute</orgName>
		</respStmt>
	</monogr>
	<note>1 Publisher</note>
</biblStruct>

<biblStruct xml:id="b38">
	<analytic>
		<title level="a" type="main">Deep Learning Based Named Entity Recognition Models for Recipes</title>
		<author>
			<persName><forename type="first">A</forename><surname>Agarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kapuriya</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Agrawal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">V</forename><surname>Konam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Goel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Rastogi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Niharika</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Bagler</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/2024.lrec-main.406" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)</title>
				<editor>
			<persName><forename type="first">N</forename><surname>Calzolari</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M.-Y</forename><surname>Kan</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">V</forename><surname>Hoste</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">A</forename><surname>Lenci</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Sakti</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">N</forename><surname>Xue</surname></persName>
		</editor>
		<meeting>the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)<address><addrLine>Torino, Italia</addrLine></address></meeting>
		<imprint>
			<publisher>ELRA and ICCL</publisher>
			<date type="published" when="2024">2024</date>
			<biblScope unit="page" from="4542" to="4554" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b39">
	<analytic>
		<title level="a" type="main">Extracting Settings from Multilingual Recipes with Various Sequence Tagging Models: an Experimental Study</title>
		<author>
			<persName><forename type="first">C</forename><surname>Radu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C.-E</forename><surname>Staicu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L.-M</forename><surname>Mitrică</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Dînşoreanu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Potolea</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Lemnaru</surname></persName>
		</author>
		<idno type="DOI">10.1109/ICCP56966.2022.10053968</idno>
		<ptr target="2766-8495" />
	</analytic>
	<monogr>
		<title level="m">IEEE 18th International Conference on Intelligent Computer Communication and Processing (ICCP)</title>
				<imprint>
			<date type="published" when="2022">2022. 2022</date>
			<biblScope unit="page" from="65" to="72" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b40">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Wróblewska</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kaliska</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Pawłowski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Wiśniewski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Sosnowski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ławrynowicz</surname></persName>
		</author>
		<ptr target="http://arxiv.org/abs/2204.07775" />
		<title level="m">TASTEset -Recipe Dataset and Food Entities Recognition Benchmark</title>
				<imprint>
			<date type="published" when="2022">2022</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b41">
	<analytic>
		<title level="a" type="main">SentencePiece: A simple and language independent subword tokenizer and detokenizer for Neural Text Processing</title>
		<author>
			<persName><forename type="first">T</forename><surname>Kudo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Richardson</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/D18-2012</idno>
		<ptr target="https://aclanthology.org/D18-2012.doi:10.18653/v1/D18-2012" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">E</forename><surname>Blanco</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">W</forename><surname>Lu</surname></persName>
		</editor>
		<meeting>the 2018 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, Association for Computational Linguistics<address><addrLine>Brussels, Belgium</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="66" to="71" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b42">
	<analytic>
		<title level="a" type="main">Hidden markov models</title>
		<author>
			<persName><forename type="first">P</forename><surname>Blunsom</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Lecture notes</title>
		<imprint>
			<biblScope unit="page">48</biblScope>
			<date type="published" when="2004-08-15">August 15 (2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b43">
	<monogr>
		<author>
			<persName><forename type="first">P</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Chen</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2111.09543</idno>
		<title level="m">Debertav3: Improving deberta using electra-style pre-training with gradient-disentangled embedding sharing</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b44">
	<analytic>
		<title level="a" type="main">Know what you don&apos;t know: Unanswerable questions for SQuAD</title>
		<author>
			<persName><forename type="first">P</forename><surname>Rajpurkar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Jia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Liang</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/P18-2124</idno>
		<ptr target="https://aclanthology.org/P18-2124.doi:10.18653/v1/P18-2124" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics</title>
				<editor>
			<persName><forename type="first">I</forename><surname>Gurevych</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Y</forename><surname>Miyao</surname></persName>
		</editor>
		<meeting>the 56th Annual Meeting of the Association for Computational Linguistics<address><addrLine>Melbourne, Australia</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page" from="784" to="789" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b45">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Schweter</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Baiter</surname></persName>
		</author>
		<ptr target="https://github.com/dbmdz/berts" />
		<title level="m">Dbmdz BERT Models</title>
				<imprint>
			<date type="published" when="2019">2019. 2024-04-22</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b46">
	<monogr>
		<title level="m" type="main">Text Chunking using Transformation-Based Learning</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">A</forename><surname>Ramshaw</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">P</forename><surname>Marcus</surname></persName>
		</author>
		<idno type="DOI">10.48550/arXiv.cmp-lg/9505040</idno>
		<idno>arXiv</idno>
		<ptr target=":cmp-lg/9505040" />
		<imprint>
			<date type="published" when="1995">1995</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b47">
	<analytic>
		<title level="a" type="main">Spanish pre-trained bert model and evaluation data</title>
		<author>
			<persName><forename type="first">J</forename><surname>Cañete</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Chaperon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Fuentes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J.-H</forename><surname>Ho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Kang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Pérez</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">PML4DC at ICLR</title>
				<imprint>
			<date type="published" when="2020">2020. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b48">
	<analytic>
		<title level="a" type="main">WikiMatrix: Mining 135M parallel sentences in 1620 language pairs from Wikipedia</title>
		<author>
			<persName><forename type="first">H</forename><surname>Schwenk</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Chaudhary</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Gong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Guzmán</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2021.eacl-main.115</idno>
		<ptr target="https://aclanthology.org/2021.eacl-main.115.doi:10.18653/v1/2021.eacl-main.115" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume</title>
		<title level="s">Association for Computational Linguistics</title>
		<editor>
			<persName><forename type="first">P</forename><surname>Merlo</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">J</forename><surname>Tiedemann</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">R</forename><surname>Tsarfaty</surname></persName>
		</editor>
		<meeting>the 16th Conference of the European Chapter of the Association for Computational Linguistics: Main Volume</meeting>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="page" from="1351" to="1361" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
