<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Hardware-efective Approaches for Skill Extraction in Job Ofers and Resumes</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Laura Vásquez-Rodríguez</string-name>
          <email>laura.vasquez@idiap.ch</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Bertrand Audrin</string-name>
          <email>bertrand.audrin@ehl.ch</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Samuel Michel</string-name>
          <email>samuel.michel@idiap.ch</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Samuele Galli</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Julneth Rogenhofer</string-name>
          <email>julneth.rogenhofer@ehl.ch</email>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jacopo Negro Cusa</string-name>
          <email>j.negrocusa@gmail.com</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Lonneke van der Plas</string-name>
          <email>lonneke.vanderplas@usi.ch</email>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Arca24.com SA</institution>
          ,
          <country country="CH">Switzerland</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>EHL Hospitality Business School, HES-SO, University of Applied Sciences and Arts Western Switzerland</institution>
          ,
          <country country="CH">Switzerland</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Idiap Research Institute</institution>
          ,
          <country country="CH">Switzerland</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>Institute of Linguistics and Language Technology, University of Malta</institution>
          ,
          <country country="MT">Malta</country>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Università della Svizzera Italiana</institution>
          ,
          <country country="CH">Switzerland</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Recent work on the automatic extraction of skills has mainly focused on job ofers and not resumes while using state-of-the-art resource-intensive methods and considerable amounts of annotated data. However, in real-life industrial contexts, the computational resources and the annotated data available can be limited, especially for resumes. In this paper, we present our experiments that use hardware-efective methods and circumvent the need for large amounts of annotated data. We experiment with various methods that vary in hardware requirements and complexity. We evaluate these systems both on public and commercial data, using gold-standard for evaluation. We find that standalone rule-based and semantic model performance on the skill extraction task is limited and variable between job ofers and resumes. However, neural models can perform competitively and be more stable, even when using small datasets, with an improvement of ∼30%. We present our experiments using minimal hardware, mostly CPU-based with less than 8 GB of RAM for rule-based and semantic methods and using GPUs for neural models with a maximum memory usage for both CPU and GPU of 24 GB, with less than 25 minutes of training time.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Human Resources</kwd>
        <kwd>Skill Extraction</kwd>
        <kwd>Recruiting</kwd>
        <kwd>Natural Language Processing</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        With the development of professional social networking
sites and job boards, job ofers get more and more
applicants. On average, an online job can get 250 applications
[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], making manual handling of applications and selection
of candidates no longer possible or practical. Partial
automation of the talent acquisition process thus seems to
be imperative, leading to cost reduction and increased
eficiency [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>The pre-screening phase is one of the most likely to be
automated to filter through large volumes of resumes. This
involves recognizing diferent types of skills, determining
their proficiency (e.g., C2 level in English), and classifying
and weighting skills according to their category (e.g., hard
versus soft skills) in relationship with a specific job ofer.</p>
      <p>In this work, we focus on the first step of this process:
extracting skills from job ofers and resumes, starting with
hard skills. To do so, organizations rely on "Applicant
Tracking Systems (ATS)" that ofer a first filter through the
resumes. Most HR professionals do not necessarily have a
clear understanding of how these ATS work, and even less
of the amount of computational resources that they require.
However, this topic is paramount not only for the large
volumes of data involved in the process of recruiting but also,
for the cost that increases with each new application. Our
motivation is the development of incremental and hybrid
approaches that should be adapted to the needs of every
organization.</p>
      <p>
        This paper is developed within the context of the SEM24
project, which seeks the introduction of NLP as a means
of guidance and support to HR specialists. In this context,
we performed a bottom-up assessment of skill extraction
methods, solely focused on hard skills as the first project
stage, considering scenarios with limited annotated data
and computational resources. We hypothesize that with
an incremental approach to the skill extraction task, it is
possible to find hardware-efective hybrid methods, which
are not only competitive in cost but also in performance.
We enumerate our contributions as follows:1
• A comparative evaluation including rule-based,
semantic, and neural models for detecting hard skills
from job ofers with publicly available and labeled
datasets and from industry-owned resumes, where
datasets tend to be more restricted.
• A manual analysis of systems outputs,
highlighting the diferences between the proposed systems,
and systematically categorizing models’ mismatches
explaining human-system discrepancies beyond to
what test sets can measure together with automatic
metrics.
• An analysis of hardware requirements, gathering
insightful information on the trade-of between the
performance and resources needed for the skill
extraction task.
The skill extraction task has already been explored for about
a decade, and yet, still not truly solved [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]. Traditionally,
1We will release our code on GitHub: https://github.com/idiap/hw_
efective_skill_extraction
rule-based methods using a taxonomy have been
predominant, either by using simple rules [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], or more formally, by
recognizing entities in the text using Named Entity
Recognition (NER) [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Taxonomies have also been involved in
the classification of resumes according to predefined jobs
or skills [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] and the search of similar terms in a weakly
supervised manner [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ].
      </p>
      <p>
        With the recent progress of neural networks, there has
been an increased interest in the automation of talent
acquisition systems using more advanced machine-learning
methods in NLP. However, these contributions are scattered
across multiple tasks, domains, and languages. At the
document level, Bhola et al. [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] proposed the classification of
job ofers with a list of relevant skills. Further, they also
explored the prediction of missing skills jointly with graph
neural networks [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]. More commonly, the skill extraction
task has been addressed as the identification and labeling
of spans (i.e., sequences of words in texts) and in diferent
languages such as English [
        <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
        ], Danish [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ] and German
[
        <xref ref-type="bibr" rid="ref13">13</xref>
        ]. The transformer architecture has also been explored,
where Zhang et al. [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] proposed fine-tuning existing
multilingual Large Language Models (LLMs) enriched with
taxonomies such as ESCO.2 More recently, papers based on
instruct-based models without further training have
proposed novel avenues for NER skill extraction [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ].
      </p>
      <p>
        Data privacy regulations introduce another relevant
challenge. Resumes are considered personal data, and they can
potentially represent ethical issues if they are not handled
correctly. Therefore, public datasets of resumes are scarce
and scattered, while job ofers are found in multiple
languages. In English, Green et al. [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ] performed the
extraction and annotation of skills and their proficiency from UK
job boards in multiple domains such as IT and finance. This
domain has also been explored in other languages such as
German [
        <xref ref-type="bibr" rid="ref16">16</xref>
        ] and French [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ]. Also, datasets have been
proposed for detecting scams in online jobs [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ] and for
the identification of privacy-related entities (e.g., names,
emails) in job postings [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. However, skills extraction from
resumes is as important as skills extraction from job ofers
in an industrial setting, or arguably even more important,
because numbers are larger and automation is key. To
mitigate these limitations, the research community has also
focused on the synthetic generation of resumes [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] and job
ofers [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ], avoiding the dificulty of handling users’ privacy
and scarcity of data.
      </p>
      <p>In this work, we challenge the assumption that
computing resources and datasets are readily available, assessing
possible and feasible scenarios using low-cost hardware
(e.g., CPU) and limited annotated data for skills extraction.
Further, we demonstrate the performance of the skill
extraction task with our proposed minimal settings, on a wide
spectrum of methods increasingly transitioning in method
complexity, resource consumption, and efectiveness. To
our knowledge, no work has explored the skill extraction
task from a resource-efective perspective.</p>
    </sec>
    <sec id="sec-2">
      <title>3. Methodology</title>
      <p>We define our proposed task in Section 3.1. We investigate
the extraction of hard skills in both scenarios from publicly
available labeled job ofers and resumes, which are more
restricted in access while using methods that require minimal
GPU. Next, we detail the data collection process in Section
2https://esco.ec.europa.eu/en
3.2, the selected extraction methods in Section 3.3, and
finally, human evaluation of the system outputs are described
in Section 3.4.</p>
      <sec id="sec-2-1">
        <title>3.1. Tasks Definition</title>
        <p>
          This paper explores extracting hard skills from job ofers and
resumes. We selected the NER task [
          <xref ref-type="bibr" rid="ref22">22</xref>
          ] to detect skills and
occupations as entities. The main diference between the
traditional NER and our approach is that we will perform
the extraction of text spans (i.e., a contiguous sequence of
words), whose length can vary greatly between entities and,
in some cases, it may not represent a single concept (e.g.,
"collaborating with multiple teams").
        </p>
        <p>We subdivide our task according to the access level (i.e.,
public or restricted) of HR datasets. In this domain, datasets
containing job ofers are often publicly available, whereas
datasets with resumes are almost non-existent, limiting the
development of extraction methods in multiple scenarios
(e.g., domains and languages). As the quality and level of
access difer between resumes and job ofers, we propose
the following tasks:
Task 1: Skill Extraction from Job Ofers: The
extraction of hard skills using annotated data is well explored with
various methods. Using human-annotated data ensures a
better quality of the outputs; however, this type of data is
not available for all languages and domains. An additional
advantage of working on job ofers is that data can be shared
without ethical concerns which favours reproducibility. In
this task, we will automatically annotate job ofers using
a taxonomy, evaluating the results with publicly available
datasets.</p>
        <p>Task 2: Skill Extraction from Resumes: There is a
scarcity of annotated data for the scenario of skill extraction
from resumes. Also, the content and the layout are more
variable, afecting the accuracy of models that are trained on
more standardized content such as job ofers. We propose
the manual annotation of resumes for training and testing
using labeled data and a manual analysis of the outputs to
determine the performance of this task.
For our experiments, we tried to satisfy both industrial and
academic requirements. We perform the proposed task in
the best possible yet cost-efective way, but we also make
sure that methods and experiments are reproducible and
shareable with the research community. To achieve this goal,
we use a combination of public and in-house, professionally
crafted taxonomies3 as a knowledge source. Second, we
selected public datasets related to job ofers and
industryowned resumes for training and testing.4</p>
        <sec id="sec-2-1-1">
          <title>3.2.1. Manually Annotated Taxonomies</title>
          <p>
            Arca24_DB: crafted by in-house professional annotators
from our industrial partner of the SEM24 project. This
taxonomy has 10,379 entities, including domains, skills, and
job occupations, available in 10 languages. We consider that
the use of this taxonomy is relevant given that skills are
extracted from actual resumes and job ofers, curated by
HR specialists. Other taxonomies such as ESCO (see below)
tend to be more standardized, with a language style
dissimilar to the ones in resumes and job ofers, representing a
challenge for skill extraction task [
            <xref ref-type="bibr" rid="ref23">23</xref>
            ]. In this work, we will
perform skill extraction from English resumes and job ofers.
Also, we focused on the selection of skills and occupations,
discarding other elements in the hierarchy (Figure 1) such
as domain entities, resulting in a total of 10,356 entities.
ESCO_DB: a more comprehensive and publicly available
taxonomy of 131,623 entities. To adapt this taxonomy to our
use case, and to make evaluations comparable, we
downloaded the ESCO dataset (v1.1.1, content classification) for
English.5 We identified all the files and columns relevant
to skills and knowledge, creating a single knowledge base
with no duplicates.
          </p>
        </sec>
        <sec id="sec-2-1-2">
          <title>3.2.2. Manually Annotated Datasets</title>
          <p>
            Green_JOB [
            <xref ref-type="bibr" rid="ref11">11</xref>
            ]: a collection of job ofers from UK job
boards annotated with skills (18,617 entities for training
and 908 for testing).6 It was annotated by crowdsourcing
with the definition of the following types of entities: skills,
knowledge, occupation, experience, and domain. As some of
these entities can be quite similar, we have remapped them
into 2 categories as follows: skills, knowledge as hard skills,
and experience and domain together with occupations. For
our experiments, we have used the train, validation, and
test splits published on HuggingFace (HF).7
Arca24_CV: we collected a set of 50 resumes in English
from multiple domains (i.e., IT, Finance, Sales). The
selection of the CVs was a random sample from a larger corpus
from Arca24_CV. The HR specialists (2 people) from our
project performed the annotations of these resumes jointly,
discussing possible disambiguation of the results. We
selected the following entities for the annotations: hard skills,
soft skills, knowledge, language, occupation, domain, and
degrees/certifications. Also, as a way of easing the
annotation process, we automatically highlighted the dates in
the text. Similarly, as in the Green_JOB dataset, we also
remapped the existing entities to achieve consistency across
3We refer to a taxonomy as a collection of skills and knowledge,
hierarchically connected to occupations in diferent domains, as shown in
Figure 1.
4In Table 1, we show the statistics of the collected and the automatically
annotated datasets, which will be explained further in Section 4.
5https://esco.ec.europa.eu/en/use-esco/download
6https://www.kaggle.com/datasets/airiddha/trainrev1
7https://huggingface.co/datasets/jjzha/green
all datasets and better performance (See Section 4.1). As for
degrees/certifications, we mapped them under hard skills.
Also, we discarded soft skills and languages in our
experiments as this will be explored in future work.8
          </p>
        </sec>
      </sec>
      <sec id="sec-2-2">
        <title>3.3. Methods in a Nutshell</title>
        <p>In this section, we give an overview of the implemented
methods (i.e., rule, semantic, and neural) for the skill
extraction task. Further, we expand on the technical details of our
methods in Section 4.2.</p>
        <sec id="sec-2-2-1">
          <title>3.3.1. Rule-based Methods</title>
          <p>The implementation of a rule-based model is
straightforward. We define a set of rules that allows the matching
of concepts in a knowledge base or taxonomy and word
sequences in a text. These concepts should be normalized,
as these could be shown in diferent surface forms (e.g.,
engineer, engineering) while referring to equally relevant
meanings for recruitment purposes.</p>
          <p>
            We performed this task by applying multiple
transformations to the concepts in the taxonomy and the word
sequences in the texts as follows: 1) Remove punctuation
and spaces through word tokenization, 2) Lemmatisation
(e.g., better → good), 3) Stemming (e.g., python
development → python develop) and 4) Expansion of the search
space by considering not only words but also subwords (e.g.,
"communicate on-line" → "communicate on-line",
"communicate", "on-line"). The normalized text from both taxonomy
and text are compared to find possible matches. As a
result, the same text span may match multiple concepts from
the taxonomy. As a final disambiguation step, candidate
concepts are selected considering their semantic similarity,9
Levenshtein distance [
            <xref ref-type="bibr" rid="ref24">24</xref>
            ] and skills length.
          </p>
          <p>The main advantage of this approach is that the output
is highly predictable, precise, explainable, and controllable
because there is a direct match between the taxonomy and
the skills extracted. The extracted skills can be mapped
to the knowledge base and skills that the system failed to
extract can be attributed to the fact that they are missing
in the taxonomy. However, keeping a taxonomy up-to-date
involves manual labor, which is time-consuming and
expensive. There are often problems with the coverage of such
hand-built resources, which leads to problems of recall in
system output.</p>
          <p>
            Another limitation of purely rule-based systems is that
they do not generalize well, as they are tied to fixed concepts.
Hence, if a concept is absent in the taxonomy, that particular
skill will not be detected. The taxonomy must be updated
often, as new jobs and skills are constantly required in the
job market [
            <xref ref-type="bibr" rid="ref25">25</xref>
            ]. Also, if similar skills are expressed with
diferent words (e.g., reporter, journalist), they will not be
detected.
          </p>
        </sec>
        <sec id="sec-2-2-2">
          <title>3.3.2. Methods based on Semantic Similarity</title>
          <p>In view of the above-mentioned limitations, we do not only
attempt to find a direct mapping between skills in the text
and concepts in the taxonomy but allow a mapping between
8For consistency, we use the term "Resume" throughout the paper,
however, for naming convenience, we use "CV" to name our dataset. Please
note that we refer to the same concept in both cases.
9Although we use semantic similarity in the final disambiguation step,
we consider the model as rule-based since the main selection of skills
is limited to the concepts that are explicitly in the taxonomy.
semantically similar terms as well. We use a pre-trained
token-based embedding model to encode the n-grams of the
original text (i.e., job ofers and resumes) and the concepts in
the taxonomy. Then, we calculate the similarity of each pair
to determine how close they are to each other. However, the
longest skills in the taxonomy can range up to eight words
in size, this leads to unreliable results as most of the skills
are smaller.</p>
          <p>To obtain a better threshold for the skill search and avoid
unnecessary comparisons, we estimated the maximum
number of tokens per skill. To achieve this, we calculated the
average size of a skill in the taxonomy, resulting in a range
of one to four words. When multiple words are present in
the taxonomy or in the text, word vector values are
averaged, meaning word matching does not depend on the token
order.10 We benefit from this approach, as smaller groups of
tokens will have less sparse representations. Finally, we
perform a selection of skills using a threshold for the semantic
similarity and Levenshtein distance, similarly as performed
in the rule-based system.11 We proposed two approaches
for matching the text and concepts in the taxonomy: 1)
comparison of all the possible n-grams of the text and taxonomy
(full) and 2) comparison of those n-grams in the text that
have not been paired yet with any skill. Hence, repetitive
comparisons of already detected texts are avoided (reduced).
The comparison of the taxonomy with all the possible text
spans is highly resource-consuming. Hence, we selected the
second approach which is faster acknowledging the
tradeof that we could get suboptimal results given that the first
match is not necessarily the best.</p>
        </sec>
        <sec id="sec-2-2-3">
          <title>3.3.3. Supervised Machine Learning Methods</title>
          <p>
            We proposed a neural, supervised setting where the
models are more likely to learn concepts and generalize better.
Although the semantic similarity methods can provide a
means of generalization between similar concepts, there is
no real learning so that the model can perform the skill
extraction task in similar domains or diferent languages
without having an explicit example for every case. Therefore, we
selected a supervised scenario where we could fine-tuned
multiple models which has no previous knowledge about
the task as BERT-base [
            <xref ref-type="bibr" rid="ref26">26</xref>
            ], but also, further models that
has domain knowledge in the field of HR such as JobBERT
[
            <xref ref-type="bibr" rid="ref10">10</xref>
            ] and ESCOXLM-R [
            <xref ref-type="bibr" rid="ref14">14</xref>
            ]. We comment further on the
technical details of these models in Section 4.2.
          </p>
        </sec>
        <sec id="sec-2-2-4">
          <title>3.3.4. Neural Methods (Instruct-based)</title>
          <p>
            Instruct-based models are also capable of performing the
skill extraction task with a NER approach. These models
are characterized to have strong inference and proficient
10https://spacy.io/usage/linguistic-features/#similarity-expectations
11In the rule-based system, we use the semantic similarity as a final step
to disambiguate matches for the same text span. The skill extraction
task is completely done by using rules to match the taxonomy.
text generation. However, the text generation is less
structured and unpredictable than the previous methods. Nguyen
et al. [
            <xref ref-type="bibr" rid="ref15">15</xref>
            ] modeled NER-style skill extraction using
GPT-3.5turbo12 model. The privacy-sensitive data we are working
with does not allow us to work with such models. Results
from this work are also not directly comparable to our
results because they are limited to the F1 scores for the Green
dataset, without details on individual precision and recall
for skills and occupations. In future work, we will
experiment with open-source instruct-based models that allow us
to work on resumes without privacy concerns.
          </p>
        </sec>
      </sec>
      <sec id="sec-2-3">
        <title>3.4. Evaluation Methods</title>
        <p>
          We performed a NER-based evaluation to assess the quality
of our task. We used the Inside-Outside-Beginning (IOB)
format [
          <xref ref-type="bibr" rid="ref27">27</xref>
          ], highlighting the text’s entities and the gold
reference using a set of predefined labels. We adapted all
systems outputs to this format to have a comparable
evaluation. The detected skills in job ofers and resumes will be
identified as entities and compared against a gold standard
(i.e., human-annotated data). We assessed our files using the
nerevaluate Python Package13 to determine the precision,
recall, and F1 score of the results for the labeled datasets
in the exact and partial evaluation. For the neural models,
we used the seqeval14 library during training for its suitable
integration with HF. However, for consistency, we report
our results using the nervaluate-based evaluation with all
model predictions exported in IOB format.
        </p>
        <p>
          For our use case in the HR domain, identifying an entity
type is challenging as each dataset has a diferent set of
entity types (e.g., skills, knowledge, domains versus hard
and soft skills). Also, it is dificult for the automatic methods
to diferentiate between similar categories (e.g., skills and
knowledge). To mitigate this scenario, we focused solely
on hard skills (e.g., Python, Inventory Management) and
occupations (e.g., Software Engineer, Sales Assistant), and
mapped other entities to these categories, as explained in
Section 4.1. Finally, we selected the exact-match schema
for detecting the skills, as proposed by Segura-Bedmar et al.
[
          <xref ref-type="bibr" rid="ref28">28</xref>
          ]. In this schema, credit is given to detected entities
regardless of the type. We also considered reporting our
results into a strict evaluation, where credits depend both
on the type and the entity. However, as for a job-resume
match the entity type is not imperative, we focused only
on the exact metric. Furthermore, we consider a
partialmatch evaluation; where there is also credit given for those
entities that are not extracted in full. This allows us to
understand which entities could potentially be detected
when the evaluation is extended beyond the defined entity
boundaries. Finally, we performed a human evaluation of
the neural outputs, which we will detail in Section 4.3.
12gpt-3.5-turbo-instruct
13https://pypi.org/project/nervaluate/
14https://github.com/chakki-works/seqeval
        </p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>4. Experiments</title>
      <p>
        In this section, we describe the datasets used, the
implementation details of our selected methods (i.e., rule-based,
semantic, and neural), the preprocessing steps for the data,
and the evaluation. Finally, we present the training details
of our models in Section 4.2.3.
4.1. Data
To perform our experiments on the Green_JOB dataset [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ],
we selected the split published in HF,15 which has been used
in previous work as well. The main diference between this
dataset and the one originally published [
        <xref ref-type="bibr" rid="ref29">29</xref>
        ], is the
redistribution of the training set to create a development set, while
the test set remains unchanged. We performed the skill
extraction task per sentence, as in the original format
provided in this dataset. We also kept the original tokenization
given by the dataset. Concerning the mapping of entities,
we mapped knowledge and qualifications together to hard
skills entities and experience to occupations entities.
      </p>
      <p>For the Arca24_CV, we annotated the dataset using
Docanno.16 We migrated the entities exported in JSONL into
the IOB format, merging spacy spans17 and doc18 with
customized alignment19 functions, as the original methods20
could not align properly all the annotated entities, showing
tokenization discrepancies with unmapped entities. This
dataset has resumes that are longer than the model input
size, hence, we divided the texts into smaller extracts. Due
to the variability of resume formats, sentence boundaries are
not always present in the text, therefore, a more structured
approach was required. To be consistent with the
sentencelevel job ofers dataset, we subdivided the Arca24_CV texts
based on the average sentence length in the Green dataset,
resulting in 23 tokens per text, split by whitespace. We also
included the restriction that no labeled entity (e.g., B-Skill,
I-Skill) should be split between sentences, hence texts could
be slightly longer.21</p>
      <sec id="sec-3-1">
        <title>4.2. Models</title>
        <p>In this section, we discuss the implementation and technical
details of our rule-based, semantic similarity, and neural
models.
15https://huggingface.co/datasets/jjzha/green
16https://github.com/doccano/doccano
17https://spacy.io/api/span
18https://spacy.io/api/doc
19https://spacy.io/api/doc#char_span
20https://spacy.io/api/top-level/#gold
21For clarity, we will also refer in our experiments to Green_JOB as "Jobs
(Green)", and to Arca24_CV as "Resumes (Ours)".</p>
        <sec id="sec-3-1-1">
          <title>4.2.1. Rule-based and Semantic</title>
          <p>For the rule-based implementation, we adapted the existing
open-source tool SkillNER.22 Also, because multiple
candidates can match the same text span, we carried out the
ifnal selection using semantic similarity with Spacy English
model.23 In the case that the word vectors were not
available, we used Levenshtein distance from NLTK package24
as an alternative. As a knowledge base for skills search, we
used Arca24_DB and ESCO_DB as taxonomies.</p>
          <p>For the semantic approach, we compared similarities
between the text and the taxonomy, using the same Spacy
models as in the rule-based model. Because semantic
similarity may propose completely unrelated skills for our use
case, we included the requirement that the candidates should
have a minimal Levenshtein distance to the concepts in the
taxonomy. The thresholds we used are 0.5 and 0.7. We
selected this lower-bound as it still shows relevant candidates
relevant to skills. As for the upper bound, we observed it
shows a more precise selection of candidates with minimal
false positives. Similarly, as in the rule-based system, we
used the same taxonomy as a reference. Concerning the
models, we used the large English Spacy model,25 which is
optimized to run in CPU.</p>
        </sec>
        <sec id="sec-3-1-2">
          <title>4.2.2. Supervised Machine Learning Methods</title>
          <p>
            Previous systems have the limitation that they rely on a
given taxonomy, where concepts are completely isolated
with no context. Hence, we experimented with the Green
and the Arca24_CV dataset to establish supervised baselines
using the available splits. For the latter, we split the dataset
using the Datasets library26 from HuggingFace (HF) [
            <xref ref-type="bibr" rid="ref30">30</xref>
            ],
resulting in train, validation and test (80/10/10) splits. We
comment further on the technical details of the proposed
models.
          </p>
          <p>
            We considered both the cased and uncased version of
BERT-base [
            <xref ref-type="bibr" rid="ref26">26</xref>
            ], a baseline model for the skill exaction task.
We selected this model as it is the base of well-established
ifne-tuned models for this task. Also, in this scenario, the
model has no previous knowledge of the NER and the skill
extraction task in the domain of HR. Previously, this 110M
of parameters model was trained for the tasks of Masked
Language Modeling (MLM) and Next Sentence Prediction
(NSP).
          </p>
          <p>
            Further, we selected JobBERT [
            <xref ref-type="bibr" rid="ref10">10</xref>
            ], a model based on the
uncased version of BERT-based, with a domain adaptive
pre-training on ∼3.2M sentences from job postings. These
22https://github.com/AnasAito/SkillNER
23https://spacy.io/models/en
24https://www.nltk.org/api/nltk.metrics.distance.html
25https://spacy.io/models/en#en_core_web_lg
26https://huggingface.co/docs/datasets/v2.21.0/en/package_reference/
main_classes#datasets.Dataset.train_test_split
models were trained for the skill extraction task,
distinguishing between skills and knowledge. Finally, we select
a large model in comparison to the selected eficient
baselines, ESCOXLM-R [
            <xref ref-type="bibr" rid="ref14">14</xref>
            ], a 559M parameters state-of-the-art
model for skill extraction, based on XLM-RoBERTa large
model. This model also performed domain-adaptive
pretraining using the available concepts in the ESCO taxonomy
27 languages. As in the previous model, there is a skill and
knowledge variant, which we also fined-tuned in the NER
task.
          </p>
          <p>With respect to the implementation details of these
models, we used HF and Pytorch Lightning libraries.27 Also,
we used the models published in HF for the BERT-based
(cased,28 uncased),29 and JobBERT (skill30 and knowledge).31
For the implementation of ESCOXLM-R model, we also
finetuned the models available for knowledge32 and skills.33 We
ifne-tuned all the models using the available train and
validation splits from the selected job ofers and resumes in
Section 3.2. Next, we tested the performance of the skill
extraction task using the held-out test split.</p>
        </sec>
        <sec id="sec-3-1-3">
          <title>4.2.3. Training Details</title>
          <p>
            In Table 5 we include the time and hardware resources
consumed for our experiments. For the neural models, we
experimented with multiple hyper-parameters as suggested
by Zhang et al. [
            <xref ref-type="bibr" rid="ref14">14</xref>
            ], including the batch size of 8, 16, 32 and
learning rate of 1− 4, 1− 5 and 5− 5. We found our best
27https://lightning.ai/docs/pytorch/stable/
28https://huggingface.co/google-bert/bert-base-cased
29https://huggingface.co/google-bert/bert-base-uncased
30https://huggingface.co/jjzha/jobbert_skill_extraction
31https://huggingface.co/jjzha/jobbert_knowledge_extraction
32https://huggingface.co/jjzha/escoxlmr_knowledge_extraction
33https://huggingface.co/jjzha/escoxlmr_skill_extraction
setting, by using a learning rate of 5− 5, batch size of 16,
and training the models for 10 epochs. In particular, for
escoxlmr_knowledge_extraction and escoxlmr_skill_extraction
we used a learning rate of 5− 5 for the Green_JOB dataset
and 1− 5 for the Arca_CV dataset, which showed a more
stable and better-performing setting for these large models.
We also seeded our experiments for reproducible results and
selected the experiments with the best-performing result
on the validation set. For our neural experiments, we report
our average results for all the seeded experiments using the
following randomly selected values: 42, 31, 22, 57, 37.
          </p>
        </sec>
      </sec>
      <sec id="sec-3-2">
        <title>4.3. Human Evaluation</title>
        <p>
          To understand the quality of our neural models, we
performed the analysis of 120 random sentences in total
for the selected models: bert-based-cased (best) and
escoxlmr_skill_extraction (larger) on the green and Arca24_CV
dataset. Following Nguyen et al. [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ], we classified the
entities in each sentence into 5 error categories: 1) Skill
definition misalignment, when the system extracts a
careerrelated term that is not a skill; 2) Wrong extraction, where
the system extracts an entity that is completely unrelated to
any skill; 3) Conjoined skills, when two skills appear in a
single text span, e.g. develop reporting software and statistical
software, but the systems sees it as one; 4) Extended span,
where the system selected entities that are longer than the
ground truth; 5) Incorrect annotations, where the human
annotation is not precise; and 6) Others. The category "Others"
included scenarios such as the incomplete detection of skills
in comparison with the ground truth. We also included an
additional category for the correct entities as well (Category
0). The evaluation was done by a domain expert on the skill
extraction task. Finally, we report the percentages for each
category in our error analysis as shown in Table 2.
        </p>
        <p>exact</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>5. Results</title>
      <p>We present our results for the rule-based and semantic
systems in Table 4. For the exact evaluation, we report the
highest values for the detection of skills in job ofers.
Overall, Semantic_09_07 models are highly precise, especially for
the detection of occupations. However, in exchange for
precision, they would have a minimal recall in comparison to
the Rule-based and Semantic_09_05. In terms of the F1-score,
the rule-based systems are the more balanced when it comes
to considering both precision and recall. The taxonomy
selected, also impacts the output of the skill detection task,
while ESCO-based results tend to be less precise, these have
a higher recall due to the size of this resource. Concerning
the evaluation of resumes, a similar trend is shown in all
the systems, but with lower values given the nature of these
texts. We will discuss in detail these issues in Section 6.</p>
      <p>Further, we report our results on the neural systems in
Table 3. Given that these models are mostly based on similar
BERT models, they are not divergent between them.
However, we can observe diferences in whether the models
consider the diference between cases (i.e., small case or upper
case) or not. For both resumes and job ofers, surprisingly,
the bert-based-cased model showed the best performance
although originally there was no knowledge of the task or
domain. As a goal to compare with a resource-consuming
scenario, we also run our datasets using the escoxlmr models,
which showed to have comparable performance to previous
models. We analyze the efect of these diferences in Tables
6 and 7.</p>
      <p>Additionally, we comment on our human evaluation.
Given that the results in Table 3 are quite close, we
considered it relevant to perform an analysis of the neural
system output. We present the results of our domain expert
evaluation in Table 2. While around 50% of the samples
were correct, we report a diverse distribution in the rest of
the error categories. There is a fair share of annotations
that show misalignment between what humans and systems
consider to be a skill, followed by extended spans (when the
system extracts a longer span than the skill itself), as well as
incorrect annotations (when the system extracts an entity
but is not demonstrated in the ground truth). Next, there are
also a few conjoined skills, when the system extracts two
skills at once (e.g., develop reporting software and statistical
software).</p>
      <p>Finally, we propose the analysis of this work from a
hardware-efective perspective, presenting our resource
consumption analysis of all our systems runs in Table 5.</p>
    </sec>
    <sec id="sec-5">
      <title>6. Discussion</title>
      <p>For rule-based systems, the main limitation is the
dependency of the systems on a manually built taxonomy. Results
are shown to be more precise in the detection of concepts,
which means that the concepts that are detected are very
likely to be true. However, we found that entities are
disconnected from their context, as the main goal is to match
a rule in the taxonomy. For example, we can find generic
sentences about organizations present in job ofers that are
identified as skills (i.e., "The company specializes in
marketing, PR, Account Handling"), although they do not directly
refer to skills. If the same sentence were to appear in a
resume, this would indeed be considered a skill. With respect
to the taxonomies selected, we also expected an incremental
diference with large datasets such as ESCO, however, when
a strict rule-based approach is followed, it is impossible to
match a given concept to similar ones that have diferent
contexts and styles in terms of the language used between
the taxonomy, job ofers, and resumes. Still, we can see
cases, such as in Tables 6 and 7, when Rule-based using
ESCO shows competitive results to the neural systems.</p>
      <p>In contrast, semantic systems may result in a more
flexible extraction of skills regarding the surface properties of
text and the taxonomy, but this would afect significantly
the precision of the skills selected. For end-customers,
precision is highly important as customers’ and HR specialists’
trust depends more on whether the system is correct or not.
Hence, an automated solution would be seen as unreliable
compared to manually selecting candidates. Driven by the
significance of precision, we proposed the use of more strict
thresholds, such as in the Semantic_09_07. In this manner,
we obtained concepts that are quite close to the ones
existing in the taxonomy, however, it resulted in a very low
recall. From an inclusive perspective, recall is very relevant,
as we would like to give equal opportunity and importance
to all candidates. Therefore, the importance of balancing the
expectations of the system, without neglecting or showing
bias towards any candidate.</p>
      <p>
        Rule-based and semantic systems have the advantage that
the origin of the detection is highly traceable to a taxonomy,
and the reason for a success or failure is straightforward.
While this has been less relevant in the past, today is an
important challenge [
        <xref ref-type="bibr" rid="ref31">31</xref>
        ] which is also materialized in
existing regulations, where high-risk AI-based applications,
such as recruiting algorithms, will consider explainability
as an essential feature.34 These approaches are also
inexpensive, with the limitation that they could be potentially
34https://eur-lex.europa.eu/eli/reg/2024/1689/oj
not scalable as their performance would be bound to the
size of a given taxonomy. This could be the case of
using a large taxonomy such as ESCO with methods such as
semantic_09_05 and semantic_09_07 (See Table 5), which
represents a base worst case of processing time. However,
it is still possible to include optimizations such as the
parallelization of search algorithms, the inclusion of more flexible
and eficient database solutions, and the simplification of
taxonomies with non-relevant terms. Overall, this is not
a trivial task, however, it is worth exploring alternatives
before choosing resource-consuming solutions.
      </p>
      <p>Regarding neural-based systems, we observed a
significant increase for all the proposed scenarios compared to
the previous rule-based and semantic systems. Also, these
models achieved equivalent performance for the exact
evaluation in both job ofers and resumes, considering their
overall average (i.e., all_f1). Interestingly, for the partial
evaluation, there is a boost in the job ofer metrics. We
hypothesize that this could be due to the proximity of the
prediction to the true label, but yet, is not as precise as the
exact annotation. Job ofers tend to be more standardized,
while resumes can express the same skill diferently. Also,
there is more noise in resumes, as data is often available
in PDF format and layouts can vary greatly between
candidates. This poses a challenge to the skill extraction task in
settings using NER, where it is expected to find a bounded
entity within the existing text. Also, these methods discard
possible skills that are not explicitly in the text, which is a
typical characteristic of soft skills.</p>
      <p>We consider that the main challenge has been the
selection of quality data. We not only account for the variability
of the resumes and the noise of data parsing but also, the
restricted access of resumes. While we partnered with a
company that can provide these data, it is mostly unlabeled for
supervised settings. Also, quality test sets are scarce which
represents a time-consuming job and challenging task for
annotators to achieve accurate labeling of skills in multiple
domains. For our use case, we selected the Green dataset as
a state-of-the-art human-annotated resource. This dataset
is publicly available, which allows the reproducibility of the
experiments. Nevertheless, this dataset represents a
collection of sentences from job ofers from the UK, showing a
limited domain and context to evaluate our systems.</p>
      <p>Further, to support the quality analysis of our data, we
comment on the manual analysis. We consider that the
major challenge of skill extraction is the subjective nature
of this task, supported by most of the annotations in the
ifrst category. We also found it interesting that although
we scored similarly on the results, the distribution of the
errors was significantly diferent. There is also a
minority related to skills that need to be inferred, but we also
acknowledge that these datasets are mostly dedicated to
explicit skill extraction. Also, the detection of entities within
rigid boundaries in tasks such as NER can be dificult, which
was clear in the errors reported for Category 4.</p>
      <p>Finally, we discuss the main topic and motivation of our
work: hardware-efective skill extraction methods. While
the NLP hype pushes toward large-scale LLMs, these are still
under consideration for real-user cases when it comes to cost
and explainability. Within the HR domain, companies may
process thousands of resumes per week, hence, representing
an increased cost directly associated with the data volume.
Also, customers would be required to understand the reason
for selection which is hard to track in non-deterministic
systems. In our work, we show the possibilities of systems
that can rely solely upon CPUs or within limited GPU
training time. We believe it is important to evaluate the benefit
between these methods and LLMs, including the budget
that needs to be invested. For example, the escoxlm-based
models represent 5X the size of the bert-based models and
more than 3X in resource consumption, however, the benefit
in detection is small and similar to the original baselines.
We would like to stress with our work, that it is valuable
to evaluate a wide spectrum of solutions that not only rely
on large models and datasets. We understand that some
LLMs could be proficient in many HR-related tasks, but we
should consider hybrid approaches that not only benefit in
precision, and cost-efective but also explainability for fair
and inclusive AI-based recruiting.</p>
    </sec>
    <sec id="sec-6">
      <title>7. Conclusions and Future Work</title>
      <p>In this paper, we have proposed three diferent skill
extraction methods, for the detection of hard skills, using multiple
configurations and datasets, including resumes and job
offers. We have exposed the boundaries of rule-based,
semantic, and neural methods using minimal hardware, including
CPU-based architectures with less than 8GB of RAM and/or
minimal GPU training in less than 25 minutes and 24GB
RAM. Also, we have analyzed the performance of these
systems by using supervised baselines, showing that the latter
can improve ∼30% with minimal data for the
taxonomybased ones.</p>
      <p>In future work, we aim to experiment with more
complex architectures, such as instruct-based LLMs and more
annotated datasets by professional experts in diferent
languages. The presented state-of-the-art models show that
these are quite competitive, however, they tend to be less
explainable to a non-technical audience (e.g. business and
HR professionals) due to their indeterministic nature.</p>
      <p>Our current experiments represent a starting point to
fully understand the correct approach for the skill extraction
task, where resource optimization and explainability are
important. Further, we will also expand our experiments in
a multilingual setting, including French, Italian, Portuguese,
Spanish, Italian, and German. Next, we will continue with
an essential part of this task which is the inclusion and
analysis of soft skills, which truly represent the human
aspect of recruiting. Also, we would investigate how to cast
the skill extraction task as a sentence classification task,
because NER is tightly linked particular literal text span.</p>
    </sec>
    <sec id="sec-7">
      <title>Ethics Statement</title>
      <p>Developing NLP algorithms in the HR domain is challenging
due to the relatively scarce availability of public datasets and
evaluation scenarios. For our experiments, we have used
public datasets to ensure the reproducibility of our methods.
However, we also include industry-proprietary datasets and
resumes with sensitive information. We acknowledge that
we have acquired the corresponding licenses and data
consents for managing this information. However, due to the
nature of these resources and the ethical considerations that
come with them, we are not able to share them publicly.</p>
    </sec>
    <sec id="sec-8">
      <title>Limitations</title>
      <p>The availability of public datasets for the task of skills
extraction is limited, where there are small samples per domain
and languages. As a start, we have relied on available
resources for English, however, these are not large and some
are specific (e.g. detection of hard skills in UK-based job
ofers). Similarly, public resumes are mostly unavailable,
which limits the possibility of benchmarking our methods
with data that we are allowed to share with no privacy
concerns. We acknowledge that larger deep-learning
architectures, including instruct-based models, could show
a better performance for the proposed task; however, it is
important to consider that the availability of GPUs is not
always a given. Finally, developing a taxonomy as a
knowledge base is essential to build an explainable system that
users trust. Although depending solely on a list of concepts
shows limited performance, it can be complemented with
additional neural methods for each scenario. With respect
to the hardware resources, we understand that the memory
consumption could also be afected by external factors
outside of our control (e.g., jobs running in the same CPU/GPU
nodes). However, our relative estimation shows in
perspective the resource usage with diferent architectures and data
sizes, which is enough for the proposal of resource-saving
strategies.</p>
    </sec>
    <sec id="sec-9">
      <title>Acknowledgments</title>
      <p>We would like to thank Alexandre Nanchen for his feedback
on this paper. We also thank Ewan Roche for his support in
enabling the eficiency calculations. Finally, we gratefully
acknowledge the support from Innosuisse (grant 104.069
IP-ICT).
The Test Consultant Automation Test Analyst will ideally be confident with Selenium
and good experience of web-based testing HTML and Javascript
The Test Consultant Automation Test Analyst will ideally be confident with Selenium
and good experience of web-based testing HTML and Javascript
The Test Consultant Automation Test Analyst will ideally be confident with Selenium
and good experience of web-based testing HTML and Javascript
The Test Consultant Automation Test Analyst will ideally be confident with Selenium
and good experience of web-based testing HTML and Javascript
The Test Consultant Automation Test Analyst will ideally be confident with Selenium
and good experience of web-based testing HTML and Javascript
The Test Consultant Automation Test Analyst will ideally be confident with Selenium
and good experience of web-based testing HTML and Javascript
The Test Consultant Automation Test Analyst will ideally be confident with Selenium
and good experience of web-based testing HTML and Javascript
The Test Consultant / Automation Test Analyst will ideally be confident with Selenium
and have good experience of web-based testing HTML and Javascript
the test consultant automation test analyst will ideally be confident with selenium and
good experience of web-based testing html and javascript
The Test Consultant Automation / Test Analyst will ideally be confident with Selenium
and good experience of web-based testing, HTML and Javascript.</p>
      <p>The Test Consultant Automation Test Analyst will ideally be confident with Selenium
and good experience of web-based testing HTML and Javascript.</p>
      <p>The Test Consultant Automation Test Analyst will ideally be confident with Selenium
and good experience of web-based testing HTML and Javascript.</p>
      <p>Example
drafting of bug fix reports, project virtualization to optimize the testing process, project
documentation, reporting development, uml design training.
drafting of bug fix reports, project virtualization to optimize the testing process, project
documentation, reporting development, uml design training.
drafting of bug fix reports, project virtualization to optimize the testing process, project
documentation, reporting development, uml design training.
drafting of bug fix reports, project virtualization to optimize the testing process, project
documentation, reporting development, uml design training.
drafting of bug fix reports, project virtualization to optimize the testing process, project
documentation, reporting development, uml design training.
draft ing of bug fix reports, project virtualization to optimize the testing process, project
documentation, reporting development, uml design training.
drafting of bug fix reports, project virtualization to optimize the testing process, project
documentation, reporting development, uml design training.
draft ing of bug fix reports, project virtualization to optimize the testing process, project
documentation, reporting development, uml design training.
draft ing of bug fix reports, project virtualization to optimize the testing process, project
documentation, reporting development, uml design training.
drafting of bug fix reports, project virtualization to optimize the testing process, project
documentation, reporting development, uml design training.
draft ing of bug fix reports, project virtualization to optimize the testing process, project
documentation, reporting development, uml design training.
draft ing of bug fix reports, project virtualization to optimize the testing process, project
documentation, reporting development, uml design training.
drafting of bug fix reports, project virtualization to optimize the testing process, project
documentation, reporting development, uml design training.
Rule-based
Semantic
bert-base_ cased
bert-base-
uncased
jobbert_ skill_
extraction
jobbert_
knowledge_ extraction
escoxlmr_ skill_
extraction
escoxlmr_
knowledge_ extraction</p>
      <p>Taxonomy /
Train
Arca24_ DB
ESCO
Arca24_ DB /
09_05
Arca24_DB /
09_07
ESCO_DB /
09_05
ESCO_DB /
09_07
Green
Green
Green
Green
Green
Taxonomy
/ Train
Arca24_ DB
ESCO
Arca24_DB
/ 09_05
Arca24_DB
/ 09_07
ESCO /
09_05
ESCO /
09_07
Resumes
(Ours)
Resumes
(Ours)
Resumes
(Ours)
Resumes
(Ours)
Resumes
(Ours)
Resumes
(Ours)</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J. S.</given-names>
            <surname>Black</surname>
          </string-name>
          , P. v. Esch,
          <article-title>AI-enabled recruiting: What is it and how should a manager use it?</article-title>
          ,
          <source>Business Horizons</source>
          <volume>63</volume>
          (
          <year>2020</year>
          )
          <fpage>215</fpage>
          -
          <lpage>226</lpage>
          . URL: https://www.sciencedirect.com/ science/article/pii/S0007681319301612. doi:https: //doi.org/10.1016/j.bushor.
          <year>2019</year>
          .
          <volume>12</volume>
          .001.
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>R. D.</given-names>
            <surname>Johnson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. L.</given-names>
            <surname>Stone</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. M. Lukaszewski</surname>
          </string-name>
          ,
          <article-title>The benefits of eHRM and AI for talent acquisition</article-title>
          ,
          <source>Journal of Tourism Futures</source>
          <volume>7</volume>
          (
          <year>2021</year>
          )
          <fpage>40</fpage>
          -
          <lpage>52</lpage>
          . URL: https://doi.org/10.1108/JTF-02-2020-0013. doi:
          <volume>10</volume>
          . 1108/JTF-02-2020-0013, publisher: Emerald Publishing Limited.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>E.</given-names>
            <surname>Senger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , R. van der Goot,
          <string-name>
            <given-names>B.</given-names>
            <surname>Plank</surname>
          </string-name>
          ,
          <article-title>Deep learning-based computational job market analysis: A survey on skill extraction and classification from job postings</article-title>
          , in: E. Hruschka,
          <string-name>
            <given-names>T.</given-names>
            <surname>Lake</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Otani</surname>
          </string-name>
          , T. Mitchell (Eds.),
          <source>Proceedings of the First Workshop on Natural Language Processing for Human Resources (NLP4HR</source>
          <year>2024</year>
          ),
          <article-title>Association for Computational Linguistics, St</article-title>
          .
          <source>Julian's, Malta</source>
          ,
          <year>2024</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>15</lpage>
          . URL: https: //aclanthology.org/
          <year>2024</year>
          .nlp4hr-
          <fpage>1</fpage>
          .1.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>K.</given-names>
            <surname>Bothmer</surname>
          </string-name>
          , T. Schlippe,
          <article-title>Skill scanner: Connecting and supporting employers, job seekers and educational institutions with an ai-based recommendation system, in: Innovative Approaches to Technology-Enhanced Learning for the Workplace</article-title>
          and Higher Education, Springer International Publishing, Cham,
          <year>2023</year>
          , pp.
          <fpage>69</fpage>
          -
          <lpage>80</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Javed</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Jacob</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>McNair, Skill: A system for skill identification and normalization</article-title>
          ,
          <source>Proceedings of the AAAI Conference on Artificial Intelligence</source>
          <volume>29</volume>
          (
          <year>2015</year>
          )
          <fpage>4012</fpage>
          -
          <lpage>4017</lpage>
          . URL: https://ojs.aaai.org/index.php/ AAAI/article/view/19064. doi:
          <volume>10</volume>
          .1609/aaai.v29i2.
          <fpage>19064</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>D. A.</given-names>
            <surname>Tamburri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.-J. V. D.</given-names>
            <surname>Heuvel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Garriga</surname>
          </string-name>
          ,
          <article-title>Dataops for societal intelligence: a data pipeline for labor market skills extraction and matching</article-title>
          ,
          <source>in: 2020 IEEE 21st International Conference on Information Reuse and Integration for Data Science (IRI)</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>391</fpage>
          -
          <lpage>394</lpage>
          . doi:
          <volume>10</volume>
          .1109/IRI49571.
          <year>2020</year>
          .
          <volume>00063</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. N.</given-names>
            <surname>Jensen</surname>
          </string-name>
          , R. van der Goot,
          <string-name>
            <given-names>B.</given-names>
            <surname>Plank</surname>
          </string-name>
          ,
          <article-title>Skill extraction from job postings using weak supervision</article-title>
          ,
          <source>in: RecSys in HR'22: The 2nd Workshop on Recommender Systems for Human Resources, in conjunction with the 16th ACM Conference on Recommender Systems, September 18-23</source>
          ,
          <year>2022</year>
          , Seattle, USA.,
          <source>CEUR Workshop Proceedings</source>
          ,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bhola</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Halder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Prasad</surname>
          </string-name>
          , M.-
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kan</surname>
          </string-name>
          ,
          <article-title>Retrieving skills from job descriptions: A language model based extreme multi-label classification framework</article-title>
          , in: D.
          <string-name>
            <surname>Scott</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Bel</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          Zong (Eds.),
          <source>Proceedings of the 28th International Conference on Computational Linguistics</source>
          ,
          <source>International Committee on Computational Linguistics</source>
          , Barcelona,
          <source>Spain (Online)</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>5832</fpage>
          -
          <lpage>5842</lpage>
          . URL: https://aclanthology.org/
          <year>2020</year>
          .coling-main.
          <volume>513</volume>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2020</year>
          .coling-main.
          <volume>513</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>N.</given-names>
            <surname>Goyal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kalra</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Mutharaju</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Sachdeva</surname>
          </string-name>
          , P. Kumaraguru,
          <article-title>JobXMLC: EXtreme multi-label classification of job skills with graph neural networks</article-title>
          , in: A.
          <string-name>
            <surname>Vlachos</surname>
          </string-name>
          , I. Augenstein (Eds.),
          <source>Findings of the Association for Computational Linguistics: EACL</source>
          <year>2023</year>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Dubrovnik, Croatia,
          <year>2023</year>
          , pp.
          <fpage>2181</fpage>
          -
          <lpage>2191</lpage>
          . URL: https://aclanthology.org/
          <year>2023</year>
          .findings-eacl.
          <volume>163</volume>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .findings-eacl.
          <volume>163</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Jensen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sonniks</surname>
          </string-name>
          ,
          <string-name>
            <surname>B. Plank,</surname>
          </string-name>
          <article-title>SkillSpan: Hard and soft skill extraction from English job postings</article-title>
          , in: M.
          <string-name>
            <surname>Carpuat</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.-C. de Marnefe</surname>
            ,
            <given-names>I. V.</given-names>
          </string-name>
          <string-name>
            <surname>Meza Ruiz</surname>
          </string-name>
          (Eds.),
          <source>Proceedings of the</source>
          <year>2022</year>
          <article-title>Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Association for Computational Linguistics</article-title>
          , Seattle, United States,
          <year>2022</year>
          , pp.
          <fpage>4962</fpage>
          -
          <lpage>4984</lpage>
          . URL: https:// aclanthology.org/
          <year>2022</year>
          .naacl-main.
          <volume>366</volume>
          . doi:
          <volume>10</volume>
          .18653/ v1/
          <year>2022</year>
          .naacl-main.
          <volume>366</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>T.</given-names>
            <surname>Green</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Maynard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <article-title>Development of a benchmark corpus to support entity recognition in job descriptions</article-title>
          , in: N.
          <string-name>
            <surname>Calzolari</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Béchet</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Blache</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Choukri</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Cieri</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Declerck</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Goggi</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Isahara</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Maegaard</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Mariani</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Mazo</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Odijk</surname>
          </string-name>
          , S. Piperidis (Eds.),
          <source>Proceedings of the Thirteenth Language Resources and Evaluation Conference</source>
          , European Language Resources Association, Marseille, France,
          <year>2022</year>
          , pp.
          <fpage>1201</fpage>
          -
          <lpage>1208</lpage>
          . URL: https://aclanthology.org/
          <year>2022</year>
          . lrec-
          <volume>1</volume>
          .
          <fpage>128</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. N.</given-names>
            <surname>Jensen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Plank</surname>
          </string-name>
          , Kompetencer:
          <article-title>Finegrained skill classification in Danish job postings via distant supervision and transfer learning</article-title>
          , in: N.
          <string-name>
            <surname>Calzolari</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Béchet</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Blache</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Choukri</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Cieri</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Declerck</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Goggi</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Isahara</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Maegaard</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Mariani</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Mazo</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Odijk</surname>
          </string-name>
          , S. Piperidis (Eds.),
          <source>Proceedings of the Thirteenth Language Resources and Evaluation Conference</source>
          , European Language Resources Association, Marseille, France,
          <year>2022</year>
          , pp.
          <fpage>436</fpage>
          -
          <lpage>447</lpage>
          . URL: https://aclanthology.org/
          <year>2022</year>
          .lrec-
          <volume>1</volume>
          .
          <fpage>46</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <surname>A</surname>
          </string-name>
          .
          <article-title>-s.</article-title>
          <string-name>
            <surname>Gnehm</surname>
            , E. Bühlmann,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Buchs</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Clematide</surname>
          </string-name>
          ,
          <article-title>Fine-grained extraction and classification of skill requirements in German-speaking job ads</article-title>
          , in: D.
          <string-name>
            <surname>Bamman</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Hovy</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Jurgens</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Keith</surname>
            ,
            <given-names>B. O</given-names>
          </string-name>
          <string-name>
            <surname>'Connor</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          Volkova (Eds.),
          <source>Proceedings of the Fifth Workshop on Natural Language Processing and Computational Social Science (NLP+CSS)</source>
          ,
          <article-title>Association for Computational Linguistics, Abu Dhabi</article-title>
          ,
          <string-name>
            <surname>UAE</surname>
          </string-name>
          ,
          <year>2022</year>
          , pp.
          <fpage>14</fpage>
          -
          <lpage>24</lpage>
          . URL: https://aclanthology.org/
          <year>2022</year>
          .nlpcss-
          <volume>1</volume>
          .2. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2022</year>
          .nlpcss-
          <volume>1</volume>
          .2.
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , R. van der Goot, B. Plank, ESCOXLMR:
          <article-title>Multilingual taxonomy-driven pre-training for the job market domain</article-title>
          , in: A.
          <string-name>
            <surname>Rogers</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Boyd-Graber</surname>
          </string-name>
          , N. Okazaki (Eds.),
          <source>Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume</source>
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>Association for Computational Linguistics</source>
          , Toronto, Canada,
          <year>2023</year>
          , pp.
          <fpage>11871</fpage>
          -
          <lpage>11890</lpage>
          . URL: https://aclanthology.org/
          <year>2023</year>
          .
          <article-title>acl-long</article-title>
          .
          <volume>662</volume>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2023</year>
          .
          <article-title>acl-long</article-title>
          .
          <volume>662</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>K.</given-names>
            <surname>Nguyen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Montariol</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Bosselut</surname>
          </string-name>
          ,
          <article-title>Rethinking skill extraction in the job market domain using large language models</article-title>
          , in: E. Hruschka,
          <string-name>
            <given-names>T.</given-names>
            <surname>Lake</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Otani</surname>
          </string-name>
          , T. Mitchell (Eds.),
          <source>Proceedings of the First Workshop on Natural Language Processing for Human Resources (NLP4HR</source>
          <year>2024</year>
          ),
          <article-title>Association for Computational Linguistics, St</article-title>
          .
          <source>Julian's, Malta</source>
          ,
          <year>2024</year>
          , pp.
          <fpage>27</fpage>
          -
          <lpage>42</lpage>
          . URL: https://aclanthology.org/
          <year>2024</year>
          .nlp4hr-
          <fpage>1</fpage>
          .3.
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <surname>A</surname>
          </string-name>
          .
          <string-name>
            <surname>-S. Gnehm</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          <string-name>
            <surname>Bühlmann</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Clematide</surname>
          </string-name>
          ,
          <article-title>Evaluation of transfer learning and domain adaptation for analyzing German-speaking job advertisements</article-title>
          , in: N.
          <string-name>
            <surname>Calzolari</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Béchet</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Blache</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Choukri</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Cieri</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Declerck</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Goggi</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Isahara</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Maegaard</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Mariani</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Mazo</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Odijk</surname>
          </string-name>
          , S. Piperidis (Eds.),
          <source>Proceedings of the Thirteenth Language Resources and Evaluation Conference</source>
          , European Language Resources Association, Marseille, France,
          <year>2022</year>
          , pp.
          <fpage>3892</fpage>
          -
          <lpage>3901</lpage>
          . URL: https://aclanthology.org/
          <year>2022</year>
          .lrec-
          <volume>1</volume>
          .
          <fpage>414</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>D.</given-names>
            <surname>Beauchemin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Laumonier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. L.</given-names>
            <surname>Ster</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yassine</surname>
          </string-name>
          ,
          <article-title>"fijo": a french insurance soft skill detection dataset</article-title>
          ,
          <year>2022</year>
          . URL: https://arxiv.org/abs/2204.05208. arXiv:
          <volume>2204</volume>
          .
          <fpage>05208</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>S.</given-names>
            <surname>Vidros</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Kolias</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Kambourakis</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Akoglu</surname>
          </string-name>
          ,
          <article-title>Automatic detection of online recruitment frauds: Characteristics, methods, and a public dataset</article-title>
          ,
          <source>Future Internet</source>
          <volume>9</volume>
          (
          <year>2017</year>
          ). URL: https://www.mdpi.com/1999-5903/9/1/6. doi:
          <volume>10</volume>
          .3390/fi9010006.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>K. N.</given-names>
            <surname>Jensen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , B. Plank,
          <article-title>De-identification of privacy-related entities in job postings</article-title>
          , in: S. Dobnik, L. Øvrelid (Eds.),
          <source>Proceedings of the 23rd Nordic Conference on Computational Linguistics (NoDaLiDa)</source>
          , Linköping University Electronic Press, Sweden, Reykjavik, Iceland (Online),
          <year>2021</year>
          , pp.
          <fpage>210</fpage>
          -
          <lpage>221</lpage>
          . URL: https://aclanthology.org/
          <year>2021</year>
          .nodalida-main.
          <volume>21</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>P.</given-names>
            <surname>Skondras</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Zervas</surname>
          </string-name>
          , G. Tzimas,
          <article-title>Generating synthetic resume data with large language models for enhanced job description classification</article-title>
          ,
          <source>Future Internet</source>
          <volume>15</volume>
          (
          <year>2023</year>
          ). URL: https://www.mdpi.com/1999-5903/ 15/11/363. doi:
          <volume>10</volume>
          .3390/fi15110363.
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>A.</given-names>
            <surname>Magron</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Montariol</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Bosselut,
          <article-title>JobSkape: A framework for generating synthetic job postings to enhance skill matching</article-title>
          , in: E. Hruschka,
          <string-name>
            <given-names>T.</given-names>
            <surname>Lake</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Otani</surname>
          </string-name>
          , T. Mitchell (Eds.),
          <source>Proceedings of the First Workshop on Natural Language Processing for Human Resources (NLP4HR</source>
          <year>2024</year>
          ),
          <article-title>Association for Computational Linguistics, St</article-title>
          .
          <source>Julian's, Malta</source>
          ,
          <year>2024</year>
          , pp.
          <fpage>43</fpage>
          -
          <lpage>58</lpage>
          . URL: https://aclanthology.org/
          <year>2024</year>
          .nlp4hr-
          <fpage>1</fpage>
          .4.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sun</surname>
          </string-name>
          , J. Han,
          <string-name>
            <given-names>C.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <article-title>A survey on deep learning for named entity recognition</article-title>
          ,
          <source>IEEE Transactions on Knowledge and Data Engineering</source>
          <volume>34</volume>
          (
          <year>2022</year>
          )
          <fpage>50</fpage>
          -
          <lpage>70</lpage>
          . doi:
          <volume>10</volume>
          .1109/TKDE.
          <year>2020</year>
          .
          <volume>2981314</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>S.</given-names>
            <surname>Neutel</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. H. de Boer</surname>
          </string-name>
          ,
          <article-title>Towards automatic ontology alignment using bert</article-title>
          .,
          <source>in: AAAI Spring Symposium: Combining Machine Learning with Knowledge Engineering</source>
          ,
          <year>2021</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>12</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>V.</given-names>
            <surname>Levenshtein</surname>
          </string-name>
          ,
          <article-title>Binary codes capable of correcting deletions, insertions, and reversals</article-title>
          ,
          <source>Soviet Physics Doklady</source>
          <volume>10</volume>
          (
          <year>1966</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <surname>Cedefop</surname>
          </string-name>
          ,
          <article-title>Challenging digital myths - First findings from Cedefop's second European skills and jobs survey</article-title>
          ,
          <source>Publications Ofice of the European Union</source>
          ,
          <year>2022</year>
          . doi:doi/10.2801/818285.
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>J.</given-names>
            <surname>Devlin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Toutanova</surname>
          </string-name>
          ,
          <article-title>BERT: pre-training of deep bidirectional transformers for language understanding</article-title>
          , CoRR abs/
          <year>1810</year>
          .04805 (
          <year>2018</year>
          ). URL: http://arxiv.org/abs/
          <year>1810</year>
          .04805. arXiv:
          <year>1810</year>
          .04805.
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>L.</given-names>
            <surname>Ramshaw</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Marcus</surname>
          </string-name>
          ,
          <article-title>Text chunking using transformation-based learning</article-title>
          ,
          <source>in: Third Workshop on Very Large Corpora</source>
          ,
          <year>1995</year>
          . URL: https://aclanthology. org/W95-0107.
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>I.</given-names>
            <surname>Segura-Bedmar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Martínez</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Herrero-Zazo, SemEval-2013 task 9 : Extraction of drug-drug interactions from biomedical texts</article-title>
          (DDIExtraction
          <year>2013</year>
          ), in: S. Manandhar, D. Yuret (Eds.),
          <source>Second Joint Conference on Lexical and Computational Semantics (*SEM)</source>
          , Volume
          <volume>2</volume>
          :
          <source>Proceedings of the Seventh International Workshop on Semantic Evaluation (SemEval</source>
          <year>2013</year>
          ),
          <article-title>Association for Computational Linguistics</article-title>
          , Atlanta, Georgia, USA,
          <year>2013</year>
          , pp.
          <fpage>341</fpage>
          -
          <lpage>350</lpage>
          . URL: https://aclanthology. org/S13-2056.
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>T. A.</given-names>
            <surname>Green</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Maynard</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <article-title>Development of a benchmark corpus to support entity recognition in job descriptions</article-title>
          ,
          <source>in: Proceedings of the 13th Conference on Language Resources and Evaluation</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>1201</fpage>
          -
          <lpage>1208</lpage>
          . URL: http://www.lrec-conf.org/ proceedings/lrec2022/pdf/2022.lrec-
          <volume>1</volume>
          .128.pdf .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>T.</given-names>
            <surname>Wolf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Debut</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Sanh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Chaumond</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Delangue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Moi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Cistac</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Rault</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Louf</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Funtowicz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Davison</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Shleifer</surname>
          </string-name>
          , P. von Platen, C. Ma,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Jernite</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Plu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. L.</given-names>
            <surname>Scao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gugger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Drame</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Lhoest</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. M.</given-names>
            <surname>Rush</surname>
          </string-name>
          ,
          <article-title>Huggingface's transformers: State-ofthe-art natural language processing</article-title>
          ,
          <year>2020</year>
          . URL: https: //arxiv.org/abs/
          <year>1910</year>
          .03771. arXiv:
          <year>1910</year>
          .03771.
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Mashayekhi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Kang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lijfijt</surname>
          </string-name>
          ,
          <string-name>
            <surname>T. De Bie</surname>
          </string-name>
          ,
          <article-title>A challenge-based survey of e-recruitment recommendation systems</article-title>
          ,
          <source>ACM Comput. Surv</source>
          .
          <volume>56</volume>
          (
          <year>2024</year>
          ). URL: https://doi.org/10.1145/3659942. doi:
          <volume>10</volume>
          .1145/ 3659942.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>