<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">IberLEF 2021 Overview: Natural Language Processing for Iberian Languages</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Julio</forename><surname>Gonzalo</surname></persName>
							<affiliation key="aff0">
								<orgName type="institution">UNED NLP and IR Research Group</orgName>
								<address>
									<settlement>Madrid</settlement>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Manuel</forename><surname>Montes-Y-Gómez</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">Optics and Electronics</orgName>
								<orgName type="institution">National Institute of Astrophysics</orgName>
								<address>
									<settlement>Puebla</settlement>
									<country key="MX">Mexico</country>
								</address>
							</affiliation>
						</author>
						<author role="corresp">
							<persName><forename type="first">Paolo</forename><surname>Rosso</surname></persName>
							<email>prosso@dsic.upv.es</email>
							<affiliation key="aff2">
								<orgName type="department">PRHLT Research Center</orgName>
								<orgName type="institution">Universitat Politècnica de València</orgName>
								<address>
									<country key="ES">Spain</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">IberLEF 2021 Overview: Natural Language Processing for Iberian Languages</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">E8F783F1AEBB9EC5D0D852D07B123165</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T00:21+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Natural Language Processing</term>
					<term>Artificial Intelligence</term>
					<term>Evaluation</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Resumen IberLEF is a comparative evaluation campaign for Natural Language Processing Systems in Spanish and other Iberian languages. Its goal is to encourage the research community to organize competitive text processing, understanding and generation tasks in order to define new research challenges and set new state-of-the-art results in those languages. This paper summarizes the evaluation activities carried out in IberLEF 2021, which included twelve tasks dealing with emotions, stance and opinions, harmful information, health-related information extraction and discovery, humor and irony, and lexical acquisition. Overall, IberLEF activities were a remarkable collective effort involving 359 researchers from 22 countries in Europe, Asia and the Americas.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>IberLEF is a comparative evaluation campaign for Natural Language Processing Systems in Spanish and other Iberian languages. Its goal is to encourage the research community to organize competitive text processing, understanding and generation tasks in order to define new research challenges and set new state-of-the-art results in those languages. This paper summarizes the evaluation activities carried out in IberLEF 2021, which included twelve tasks dealing with emotions, stance and opinions, harmful information, health-related information extraction and discovery, humor and irony, and lexical acquisition. Overall, IberLEF activities were a remarkable collective effort involving 359 researchers from 22 countries in Europe, Asia and the Americas. Papers with system descriptions are included in the IberLEF 2021 Proceedings <ref type="bibr" target="#b10">[11]</ref>, and papers with task overviews have been published in the journal Procesamiento del Lenguaje Natural, vol. 67 (September 2021 issue).</p><p>In this paper we summarize the activities carried on in IberLEF 2021, extracting some aggregated figures for a better understanding of this collective effort.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">IberLEF 2021 Tasks</head><p>These are the twelve tasks organized succesfully in 2021, grouped thematically:</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Emotions, Stance and Opinions</head><p>EmoEvalEs <ref type="bibr" target="#b3">[4]</ref> was an emotion classification task, where systems were asked to predict which emotions are present in texts written in Spanish (from this set: anger, disgust, fear, joy, sadness, surprise, others). Twitter was used as textual source, and the dataset consists of 8232 manually annotated tweets. 15 research groups submitted runs for this task, out of which 11 submitted papers to the proceedings.</p><p>REST-MEX <ref type="bibr" target="#b4">[5]</ref> was an evaluation exercise focused on recommendation tasks using TripAdvisor as textual source, with texts written in several variants of Spanish (Mexican Spanish being the most common). Task 1 (Recommendation) consists in predicting the degree of satisfaction (in a 1-5 scale) of a tourist visiting a given Mexican place, given the information available in TripAdvisor about the tourist and about the site. The tourist profile includes gender, place of origin, her textual self-description in TripAdvisor, and her opinions on places she has visited. The information about the place is a brief textual description and a series of representative characteristics of the place for touristic purposes (adventure, beach, family atmosphere, etc.). Task 2 (Sentiment Polarity) consists of predicting the polarity (in a 1-5 scale) of a given TripAdvisor opinion.</p><p>Overall, the dataset gathers 2263 instances tourist/destination for the first task and 7413 opinions for the second task. 2 groups submitted results for task 1 and 7 for task 2.</p><p>VaxxStance <ref type="bibr" target="#b0">[1]</ref> focused on predicting the stance of short texts (tweets) with respect to vaccines (in favour, neutral or against). This was a multilingual task including Spanish (2697) and Basque (1384) tweets.</p><p>The challenge was addressed in three variants: in Task 1 (close track), systems could only use the text of the tweets; in Task 2 (open track), systems could use any kind of data (including tweets' metadata); finally, Task 3 (zero-shot track) was a cross-lingual stance detection challenge: systems were trained on one of the languages and tested on the other language. Three groups participated in the first task, and one in the second and third tasks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Harmful Information</head><p>There were four challenges around harmful textual information in 2021:</p><p>MeOffendES <ref type="bibr" target="#b2">[3]</ref> focused on offensive language detection in Spanish, and included two subtasks on a dataset of generic Spanish and two subtasks on a Mexican Spanish corpus. The generic Spanish dataset (OffendES) comprises 30,416 comments collected from Twitter, Instagram and Youtube; the Mexican Spanish dataset (OffendMEX) comprises 7319 annotated tweets.</p><p>The tasks on generic Spanish asked systems to predict the right class from OFP (offensive, target person), OFG (offensive, target group), OFO (offensive, target others), NOE (non offensive, but with expletive language), NO (not offensive). Systems were also asked to predict the strenght of the class, taken as the ratio of annotators than concur on the class. Subtask 1 allowed textual data as input, and Subtask 2 allowed metadata as additional input. Four teams submitted results for the first task, and one for the second.</p><p>The tasks on Mexican Spanish asked systems to do a binary prediction (offensive / not offensive), using only textual input (subtask 3) or also metadata (subtask 4). 10 groups submitted results to subtask 3 and one to subtask 4.</p><p>EXIST <ref type="bibr" target="#b13">[14]</ref> focused on the identification of sexism in Spanish and English texts, asking systems to predict whether a text has sexist content (Subtask 1) and to identify the type of sexism (ideological and inequality / stereotyping and dominance / objectification / sexual violence / mysogyny and non-sexual violence) in Subtask 2. The dataset comprises 13,000 tweets and 982 gabs. 31 groups submitted results for the first subtask, and 27 for the second.</p><p>DETOXIS <ref type="bibr" target="#b14">[15]</ref> focused on the identification of toxic content in texts, and prepared a dataset with 4359 comments from news and online forums, annotated with their level of toxicity (in a scale from 0 to 3). Subtask 1 required a binary classification (toxic / non toxic) and Subtask 2 asked systems to predict the level of toxicity in the same scale that was annotated. 31 groups submitted to the first task and 24 to the second.</p><p>Finally, FakeDeS <ref type="bibr" target="#b7">[8]</ref> focused on discovering fake news written in Spanish, and prepared a dataset with 971 news articles written in Spanish from Spain and Mexico. It was designed as a binary classification task (fake or real), and 16 groups submitted results.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Health-Related Information Extraction and Discovery</head><p>Health-Related content received special attention in IberLEF 2021, as in previous editions, with two tasks related to the medical domain:</p><p>e-HealthKD <ref type="bibr" target="#b11">[12]</ref> focused on entity recognition and classification. Systems had to recognize and classify concepts, actions, predicates and references in subtask 1, and to extract relations between them (subtask B). e-HealthKD also contemplated a main, complex task where both entity recognition and relation extraction were evaluated jointly. 8 participants submitted results to subtask A and, out of them, 7 also submitted results to subtask B and to the main challenge. The organizers performed an exhaustive annotation of 1,800 sentences extracted from MedLinePlus, WikiNews and the CORD-19 corpus.</p><p>MEDDOPROF <ref type="bibr" target="#b8">[9]</ref> worked on clinical cases (the annotations include 1844 cases extracted from medical literature), and asked systems to annotate informa-tion related to occupations/professions. Task 1 (NER) was about finding mentions of occupations and classifying each of them as a profession, an employment status or an activity; Task 2 (CLASS) involved finding mentions of occupations and determining whether they are related to the patient, to a family member, to a health professional or to someone else; and Task 3 (NORM) was about mapping predictions to one of the codes in a list of unique concept identifiers from the European Skills, Competences, Qualifications and Occupations (ESCO) classification and relevant SNOMED-CT terms. 15 groups submitted results to Task 1, 11 to Task 2 and 8 to Task 3.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.4.">Humour and Irony</head><p>There were two tasks related to Humour and Irony in 2021: HAHA <ref type="bibr" target="#b5">[6]</ref> dealt with humour detection and characterization in Spanish texts, and included four subtasks: (1) humour detection, which required determining whether a tweet was humorous or not; (2) funniness score prediction, in a 1-5 scale; (3) humour mechanism classification, out of a set of classes such as irony, wordplay, hyperbole or shock; (4) humour content classification: predict the content of the joke from a set of classes such as racist jokes, sexist jokes, dark humour, dirty jokes, etc. The dataset included 36,000 annotated tweets. 14 groups submitted to the first task, 11 to the second, 9 to the third and 8 to the fourth.</p><p>IDPT <ref type="bibr" target="#b6">[7]</ref> was a task on irony detection in Portuguese texts, defined as a binary classification problem (is this text ironic or not?). The dataset included 18494 news pieces and 15212 tweets, and 7 groups submitted results for the task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.5.">Lexical Acquisition</head><p>ADoBo <ref type="bibr" target="#b9">[10]</ref> focused on the acquisition of borrowings into Spanish from other languages (English primarily). Systems were asked to detect expressions (in Spanish news articles) that have been imported from other languages in their raw form. The dataset is an annotated collection of news articles that comprise 372,701 tokens. Four systems submitted results for this task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Aggregated Analysis of IberLEF 2021 Tasks</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Tasks characterization</head><p>In terms of languages, the distribution per tasks (including subtasks) is shown in Figure <ref type="figure" target="#fig_0">1</ref>. 74 % of the tasks deal at least with Spanish, which is the predominant subject of study in IberLEF. In terms of variants of Spanish, Spain and Mexico are the best represented, with other variants having only anecdotal presence. English is used (never as the main language) in 14 % of the tasks, and this year Basque appears for the first time in IberLEF being present in 9 % of the tasks (all belonging to VaxxStance). Finally, there is also one task dealing with Portuguese. The trend in the number of languages is positive: there were two in IberLEF 2019 (Spanish and Portuguese), only one in 2020 (Spanish) and four languages in 2021.</p><p>In terms of abstract task types, the distribution of tasks can be seen in Figure <ref type="figure" target="#fig_1">2</ref>. Out of a total of 29 tasks (each subtask is counted as a task here), 7 (24 %) are binary classification tasks, which is the most popular choice. Multiclass classification problems are also well represented with 6 tasks. There are also four tasks where classes are ordinals (e.g. 0,1,2,3) that can be interpreted either as a regression or a multiclass classification problem (regression / multiclass classification in the figure). Another variant of classification problems is ordinal classification, where classes have a relative ordering (e.g. in favour, neutral or against in stance classification): 3 tasks match this abstract task type. Finally, there is also a normalization task which implies matching profession descriptions in text with standard thesauri / ontologies, which can be seen as an extreme classification task (i.e. a classification problem where the number of classes is extremely large).</p><p>There are only 3 sequence labelling tasks, which is perhaps less than expected for an evaluation campaigned focused heavily on Natural Language Problems: tasks that identify specific structures or text chunks in text, such as named entities, fall into this category. Two of them are related to the medical domain, and the other one looks for lexical borrowings (imports from other languages).</p><p>Finally, there are two genuine regression tasks, where systems must predict a real number, and only one complex task, where the organizers try to measure the joint performance of systems in two subtasks that build together with a common goal: the e-healthKD main task. Overall, IberLEF 2021 tasks address a representative sample of abstract task types, covering a wide range of problems. Probably, to get nearer industry needs, in the future we should investigate more how to evaluate complex, end user tasks. IberLEF is also missing tasks that involve text generation, such as text summarization or machine translation problems; and tasks that involve interaction with the users, such as dialogue systems. Finally, we would like to see more application domains in the list of tasks.</p><p>In terms of evaluation metrics, the distribution can be seen in Figure <ref type="figure" target="#fig_2">3</ref>. As in previous years, there is a remarkable predominance of F1 (20 tasks used it as the main evaluation metric to rank systems), which is used for all types of classification tasks (even if it does not perfectly match the problem at hand, as in ordinal classification problems) and for sequence labelling problems. Accuracy is used in a couple of classification tasks, and Bacc (Balanced Accuracy) in another. Finally, CEM (Closeness Evaluation Metric), a metric specially useful for ordinal classification tasks and introduced recently [2] is used for one of the classification/regression tasks. Tasks interpreted as regression problems are evaluated with MSE (Mean Squared Error), RMSE (Root Mean Squared Error) and MAE (Mean Absolute Error), with none of these metrics particularly favoured.</p><p>Overall, we take this as a hint that the field might be relying too much on F1. It has some desirable properties (particularly, it is robust to the characteristics of the dataset), but it has severe limitations too. Its primary shortcoming is that it hides the actual behaviour of systems, as with all averages (F1 is a harmonic average of precision and recall). For multiclass classification, the most common procedure is to compute the arithmetic average of the harmonic averages of precision/recall across classes, which is a way of focusing exclusively on system ranking and giving up on understanding why systems fail and when. We think that the usage of F1 should be accompanied with other metrics.</p><p>Most importantly, the choice of metrics does not seem to be made justified on how the system output is going to be used, but rather on mere popularity of the metrics. This is not a shortcoming of IberLEF tasks only: most NLP challenges suffer from the same problems.</p><p>Figure <ref type="figure" target="#fig_3">4</ref> shows how IberLEF tasks have evolved in the three years that it has been running on. The number of tasks has increased (from 9 in 2019 to 12 in 2021); and in 2021 the number of new tasks is 9 (75 %), a sign that the scope of problems being studied becomes larger every year. The lower figures in 2020 are due to the irruption of COVID-19: some of the tasks could not be completed and are not depicted in the graph.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Datasets and results</head><p>In terms of types of textual sources, Figure <ref type="figure" target="#fig_4">5</ref> shows how they are used in IberLEF 2021 tasks. Twitter is the most popular source, with 15 tasks relying solely or partially on Twitter data. This does not necessarily mean that the field is primarily interested in microblogging communication; it probably reflects that collecting Twitter data is more cost effective given IPR issues and other difficulties in gathering data to redistribute to the scientific community <ref type="bibr" target="#b12">[13]</ref>. All other sources are used by at most three tasks. The good news is that there are many   In terms of dataset sizes and annotation efforts, it is difficult to establish fair comparisons, because of the diversity of text sizes and the wide variance in terms of annotation difficulty. Figure <ref type="figure" target="#fig_5">6</ref> compares dataset sizes for the classification tasks, where it is more reasonable to establish direct comparisons.</p><p>Overall the annotation effort in IberLEF 2021 is remarkable, and it is a significant contribution to enlarge test collections at least for Spanish; and, therefore, to enable significant advances in our field for Spanish and the other languages involved. The number of documents varies substantially, from over 35,000 tweets (HAHA dataset on humour) to 971 news stories for FakeDes (fake news detection). But again, direct comparisons are not fair: for instance, in the case of HAHA, they are expanding annotations on a previously existing dataset (developed in other HAHA editions); and, on the other hand, establishing whether a piece of news is fake or real is probably much more time consuming than classifying humor in tweets.</p><p>IberLEF 2021 has been carried out without funding sources (other than those obtained individually by the teams organizing and participating in the tasks). If the IberLEF organization could directly fund the task organizers, this would probably help reaching large and high quality annotations for all of the tasks accepted each year. In terms of progress with respect the state of the art, it is really difficult to extract aggregated conclusions for the whole IberLEF effort. In Figure <ref type="figure" target="#fig_6">7</ref> we display a pairwise comparison between the best system and the best baseline, for each of the tasks where at least one baseline is provided, and with respect to the official ranking metric used in each task. To avoid confusion, we have restricted the chart to tasks where the official metric varies between 0 (worst quality) and 1 (perfect output). Still, it is difficult to extract conclusions, because the effort put by task organizers in providing state-of-the-art baselines varies considerably between tasks. We can say, however, that in few cases improving the baseline has proved to be challenging, and there is one case (MeOffendEs subtask 4) where the baseline beats the best system (by a narrow margin). It would probably beneficial for future IberLEF editions to establish some minimum guidelines about the types of baselines to expect in every task; again, this would be easier to implement with dedicated funding.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Participation</head><p>Given that IberLEF 2021 was not a funded initiative, participation has been impressive, with a large fraction of current research groups interested in NLP for Spanish organizing and/or participating in one or more tasks. Overall, 359 researchers representing 173 research groups from 22 countries in Europe, Asia and the Americas were involved in IberLEF tasks. Figure <ref type="figure" target="#fig_7">8</ref> shows the distribution of research groups per country. Unsurprisingly, Spain has the largest representation, with 54 groups (note that all figures reporting participation do not collapse duplicates: a group or a researcher participating in two tasks is counted twice). Mexico has also a remarkable participation, with 36 groups. And there is a suprisingly large number of countries where no Iberian language is used but have at least two groups representing them: China (23 groups), India <ref type="bibr" target="#b9">(10)</ref>, UK (6, but note that some tasks do use English as an additional source language), Italy (4), Norway (3) and Germany (3).</p><p>Figure <ref type="figure" target="#fig_8">9</ref> shows the distribution of researchers (appearing as authors in the working notes) per country. The numbers are consistent with the distribution of groups per country, with Spain, Mexico, China, Brazil and India (the top five) representing roughly 80 % of the researchers involved. The fact that two countries in the top five, China and India, appear in the top five indicates two things: first, that Spanish attracts the attention of the NLP community at large; and second, that current NLP technologies enable processing dataset without language-specific machinery, other than pretrained language models made available to the research community. The distribution of research groups per task is shown in Figure <ref type="figure" target="#fig_9">10</ref>. Participation ranges between 31 groups (EXIST subtask 1 and DETOXIS subtask 1) and one group (MeOffendES subtask 2, VaxxStance zero-shot track and VaxxStance open track). As in other evaluation initiatives, participation seems to be driven not only by the task intrinsic interest, but also by the cost of entry: in general, classification tasks (the most basic machine learning task, for which more plug and play software packages exist) receive more participation than tasks which require more elaborated approaches and more creativity to assemble algorithmic solutions. In the middle of the table we can find most tasks in the medical domain, which attract many groups in spite of being (in general) highly challenging.</p><p>Figure <ref type="figure" target="#fig_10">11</ref> shows how participation has evolved in time; while 2020 was a difficult year with the irruption of COVID-19, in 2021 participation has grown considerably, with 173 groups (three times larger than in 2020 and a 30 % increase with respect to 2019). The number of countries involved has also grown from 18 to 22. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Conclusions</head><p>In its third edition, IberLEF has again been a remarkable collective effort for the advancement of Natural Language Processing in Spanish and other Iberian languages: with 12 main tasks and 359 researchers involved, from institutions in 22 countries in Europe, Asia and the Americas. IberLEF 2021 has been the largest up to date, and has contributed to advance the field in the areas of emotions, stance and opinions, harmful information, health-related information extraction and discovery, humour and irony, and lexical acquisition. In a field where machine learning is the ubiquitous approach to solve challenges, the definition of research challenges, their associated evaluation methodologies and the development of high-quality test collections that allow for iterative evaluation is probably the most critical step towards success. We believe IberLEF is making a significant contribution in this direction.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figura 1 .</head><label>1</label><figDesc>Figura 1. Distribution of languages in IberLEF 2021 tasks</figDesc><graphic coords="5,134.77,134.76,345.83,208.58" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figura 2 .</head><label>2</label><figDesc>Figura 2. Distribution of IberLEF 2021 tasks per abstract task type.</figDesc><graphic coords="6,134.77,134.77,345.82,238.95" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figura 3 .</head><label>3</label><figDesc>Figura 3. Distribution of official evaluation metrics in IberLEF 2021 tasks.</figDesc><graphic coords="7,134.77,134.77,345.82,204.84" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figura 4 .</head><label>4</label><figDesc>Figura 4. Evolution of IberLEF tasks across time.</figDesc><graphic coords="8,134.77,143.75,345.83,208.58" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figura 5 .</head><label>5</label><figDesc>Figura 5. Types of textual sources in IberLEF 2021 tasks.</figDesc><graphic coords="8,134.77,397.20,345.84,256.74" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figura 6 .</head><label>6</label><figDesc>Figura 6. Dataset sizes in IberLEF 2021 classification tasks.</figDesc><graphic coords="9,134.77,197.52,345.82,232.87" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Figura 7 .</head><label>7</label><figDesc>Figura 7. Performance of best systems versus baselines in IberLEF 2021 classification tasks.</figDesc><graphic coords="10,134.77,192.89,345.83,339.31" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head>Figura 8 .</head><label>8</label><figDesc>Figura 8. Number of groups participating in IberLEF 2021 tasks per country</figDesc><graphic coords="11,134.77,312.74,345.83,196.03" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_8"><head>Figura 9 .</head><label>9</label><figDesc>Figura 9. Number of researchers participating in IberLEF 2021 tasks per country.</figDesc><graphic coords="12,134.77,134.77,345.83,157.55" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_9"><head>Figura 10 .</head><label>10</label><figDesc>Figura 10. Distribution of participants per task in IberLEF 2021.</figDesc><graphic coords="12,134.77,451.44,345.81,211.49" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_10"><head>Figura 11 .</head><label>11</label><figDesc>Figura 11. Number of research groups participating in IberLEF across time.</figDesc><graphic coords="13,134.77,347.85,345.83,208.58" type="bitmap" /></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgements</head><p>The authors of this overview have been supported by the Spanish Government, Ministry of Science and Innovation, via research grants MISMIS (PGC2018-096212-B), MISMIS-BIAS (PGC2018-096212-B-C32) and MISMISFAKEnHA-TE (PGC2018-096212-B-C31); and by CONACyT-Mexico project CB-2015-01-257383 and the thematic networks program (Language Technologies Thematic Network).</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Vaxxstance@iberlef 2021: Overview of the task on going beyond text in cross-lingual stance detection</title>
		<author>
			<persName><forename type="first">R</forename><surname>Agerri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Centeno</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Espinosa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">F</forename><surname>De Landa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rodrigo</forename><surname>Álvaro</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Procesamiento del Lenguaje Natural</title>
		<imprint>
			<biblScope unit="volume">67</biblScope>
			<biblScope unit="page" from="173" to="181" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">An effectiveness metric for ordinal classification: Formal properties and experimental results</title>
		<author>
			<persName><forename type="first">E</forename><surname>Amigo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Mizzaro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Carrillo-De Albornoz</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2020.acl-main.363</idno>
		<ptr target="https://aclanthology.org/2020.acl-main.363" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics</title>
				<meeting>the 58th Annual Meeting of the Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2020-07">Jul 2020</date>
			<biblScope unit="page" from="3938" to="3949" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title level="m" type="main">Overview of meoffendes at iberlef 2021: Offensive language detection in spanish variants</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">M P</forename><surname>Del Arco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Casavantes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><forename type="middle">J</forename><surname>Escalante</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T</forename><surname>Martín-Valdivia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Montejo-Ráez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">M</forename><surname>Gómez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Jarquín-Vásquez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Villaseñor-Pineda</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="volume">67</biblScope>
			<biblScope unit="page" from="183" to="194" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Overview of the emoevales task on emotion detection for spanish at iberlef 2021</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">M P</forename><surname>Del Arco</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Jiménez-Zafra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Montejo-Ráez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">D</forename><surname>Molina-González</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Alfonso Ureña-López</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">T M V</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Procesamiento del Lenguaje Natural</title>
		<imprint>
			<biblScope unit="volume">67</biblScope>
			<biblScope unit="page" from="155" to="161" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<monogr>
		<title level="m" type="main">Overview of rest-mex at iberlef 2021: Recommendation system for text mexican tourism</title>
		<author>
			<persName><forename type="first">Álvarez</forename><surname>Carmona</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Aranda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Arce-Cardenas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Fajardo-Delgado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Guerrero-Rodríguez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>López-Monroy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">P</forename><surname>Martínez-Miranda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Pérez-Espinosa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Rodríguez-González</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">Y</forename></persName>
		</author>
		<imprint>
			<date type="published" when="2021">2021</date>
			<biblScope unit="volume">67</biblScope>
			<biblScope unit="page" from="163" to="172" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Overview of haha at iberlef 2021: Detecting, rating and analyzing humor in spanish</title>
		<author>
			<persName><forename type="first">L</forename><surname>Chiruzzo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Castro</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Góngora</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Rosá</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Meaney</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Mihalcea</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Procesamiento del Lenguaje Natural</title>
		<imprint>
			<biblScope unit="volume">67</biblScope>
			<biblScope unit="page" from="257" to="268" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Overview of the idpt task on irony detection in portuguese at iberlef 2021</title>
		<author>
			<persName><forename type="first">U</forename><forename type="middle">B</forename><surname>Corrêa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Coelho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Santos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">A</forename><surname>De Freitas</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Procesamiento del Lenguaje Natural</title>
		<imprint>
			<biblScope unit="volume">67</biblScope>
			<biblScope unit="page" from="269" to="276" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Overview of fakedes at iberlef 2021: Fake news detection in spanish shared task</title>
		<author>
			<persName><forename type="first">H</forename><surname>Gómez-Adorno</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">P</forename><surname>Posadas-Durán</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">B</forename><surname>Enguix</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Porto</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Procesamiento del Lenguaje Natural</title>
		<imprint>
			<biblScope unit="volume">67</biblScope>
			<biblScope unit="page" from="223" to="231" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Nlp applied to occupational health: Meddoprof shared task at iberlef 2021 on automatic recognition, classiffication and normalization of professions and occupations from medical texts</title>
		<author>
			<persName><forename type="first">S</forename><surname>Lima-López</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Farré-Maduell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Miranda-Escalada</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Brivá-Iglesias</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Krallinger</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Procesamiento del Lenguaje Natural</title>
		<imprint>
			<biblScope unit="volume">67</biblScope>
			<biblScope unit="page" from="243" to="256" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Overview of adobo 2021: Automatic detection of unassimilated borrowings in the spanish press</title>
		<author>
			<persName><forename type="first">E</forename><surname>Álvarez Mellado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">E</forename><surname>Anke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">G</forename><surname>Arroyo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Lignos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">P</forename><surname>Zamorano</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Procesamiento del Lenguaje Natural</title>
		<imprint>
			<biblScope unit="volume">67</biblScope>
			<biblScope unit="page" from="277" to="285" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<author>
			<persName><forename type="first">M</forename><surname>Montes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Aragón</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Agerri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Álvarez-Carmona</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Álvarez Mellado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Carrillo-De Albornoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Chiruzzo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Freitas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Gómez Adorno</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Gutiérrez</surname></persName>
		</author>
		<title level="m">Proceedings of the Iberian Languages Evaluation Forum</title>
				<editor>
			<persName><forename type="first">S</forename><forename type="middle">M</forename><surname>Jiménez-Zafra</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">S</forename><surname>Lima</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">F</forename><forename type="middle">M</forename><surname>Plaza-Del-Arco</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">M</forename><surname>Taulé</surname></persName>
		</editor>
		<meeting>the Iberian Languages Evaluation Forum<address><addrLine>IberLEF</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2021">2021. 2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Overview of the ehealth knowledge discovery challenge at iberlef 2021</title>
		<author>
			<persName><forename type="first">A</forename><surname>Piad-Morffis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Estevez-Velarde</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Gutierrez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Almeida-Cruz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Montoyo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Muñoz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Procesamiento del Lenguaje Natural</title>
		<imprint>
			<biblScope unit="volume">67</biblScope>
			<biblScope unit="page" from="233" to="242" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">On the implications of the general data protection regulation on the organisation of evaluation tasks</title>
		<author>
			<persName><forename type="first">F</forename><surname>Rangel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<ptr target="https://ojs.letras.up.pt/index.php/LLLD/article/view/6119" />
	</analytic>
	<monogr>
		<title level="j">Language and Law / Linguagem e Direito</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="80" to="102" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Overview of exist 2021: sexism identification in social networks</title>
		<author>
			<persName><forename type="first">F</forename><surname>Ródriguez-Sánchez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">C</forename><surname>De Albornoz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Plaza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gonzalo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Comet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Donoso</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Procesamiento del Lenguaje Natural</title>
		<imprint>
			<biblScope unit="volume">67</biblScope>
			<biblScope unit="page" from="195" to="207" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Overview of detoxis at iberlef 2021: Detection of toxicity in comments in spanish</title>
		<author>
			<persName><forename type="first">M</forename><surname>Taulé</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Ariza</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Nofre</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Amigó</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Rosso</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Procesamiento del Lenguaje Natural</title>
		<imprint>
			<biblScope unit="volume">67</biblScope>
			<biblScope unit="page" from="209" to="221" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
