<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>Knowledge Base Construction from Pre-trained Language Models workshop at ISWC</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Ontology Learning for ESCO: Leveraging LLMs to Navigate Labor Dynamics</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Jarno Vrolijk</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Victor Poslavsky</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Thijmen Bijl</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Maksim Popov</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Rana Mahdavi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Mohammad Shokri</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Randstad</institution>
          ,
          <addr-line>Diemermere 25, 1112TC Diemen</addr-line>
          ,
          <country country="NL">Netherlands</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>University of Amsterdam</institution>
          ,
          <addr-line>Plantage Muidergracht 12, 1018TV Amsterdam</addr-line>
          ,
          <country country="NL">Netherlands</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <volume>2024</volume>
      <fpage>0000</fpage>
      <lpage>0003</lpage>
      <abstract>
        <p>The labor market is a dynamic environment that supports numerous knowledge-driven applications through ontologies, such as ESCO and O*NET. Maintaining the relevance and accuracy of information within these ontologies and taxonomies is both resource-intensive and time-consuming. In this paper, we propose an ontology learning system that utilizes self-supervised learning, retrieval-augmented generation, and autoregressive language models to identify, classify, and link labor market mentions and entities from raw job postings. Additionally, we demonstrate the language model's ability to discover "alternative labels" and "preferred labels", and perform relation classification.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        Labor market ontologies enable the organization of information about jobs, skills, and
qualifications, facilitating communication between job seekers and employers [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. However, the
labor market is a constantly evolving environment influenced by technological advancements,
increasing individual choices, and shifting demographics. Consequently, educators, job seekers,
and lifelong learners struggle to identify the relevant knowledge, skills, abilities, and
competencies needed to distinguish themselves, each with unique objectives. Keeping these individuals
and organizations informed about labor market developments in a timely and accurate manner
is challenging and requires significant time and resources.
      </p>
      <p>
        While many knowledge-driven applications, such as ESCO [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and the O*NET [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], have proven
valuable in addressing some of the challenges within the labor market, they struggle to keep the
information in their ontologies and taxonomies relevant and up-to-date [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ]. These ontologies
provide information about occupations, knowledge, skills, competences, and qualifications.
Constructing these systems is complex, and current approaches are inadequate in handling the
incomplete and dynamic nature of real-world knowledge graphs (KGs) [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ]. These approaches
often fail to represent unseen entities, overlook the abundant textual information in ontologies
and taxonomies, and are frequently based on ontological commitments that render them
taskspecific [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>
        Extensive research has been conducted on the (semi-)automated identification of terms,
types, relations, and potential axioms from text, a process known as Ontology Learning (OL)
[
        <xref ref-type="bibr" rid="ref10 ref11 ref7 ref8 ref9">7, 8, 9, 10, 11</xref>
        ]. Traditional methods for semi-automated extraction rely on lexico-syntactic
pattern mining and clustering. However, considering the individual stages of OL—i) mention
extraction (ME) and term typing (TT), ii) the discovery of hierarchical relationships, and iii) the
discovery of non-taxonomic relationships—the recent advancements in large language models
(LLMs) ofer a cost-efective and scalable solution to OL.
      </p>
      <p>
        LLMs enable the development of general-purpose and adaptable language models that can be
tailored to various natural language processing (NLP) tasks such as classification, generation, and
sequence labeling. Adapting LLMs to specific NLP tasks involves two phases: the pre-training
phase, to obtain pre-trained language models (PLMs) typically formalized as a cloze-style task
(i.e., sequential and/or masked language models), and the downstream phase, which involves
ifne-tuning the model or prompt tuning [
        <xref ref-type="bibr" rid="ref12 ref13">12, 13</xref>
        ]. In the downstream phase, KGs are considered
in recent research to adapt PLMs to tasks such as Named Entity Recognition (NER), Relation
Extraction (RE), Open Information Extraction, Entity Linking (EL), and Relation Linking [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ].
To perform these tasks, the PLM is guided by the KG provided by the ontology (i.e., concepts,
relations, domain/range constraints) and a set of sentences.
      </p>
      <p>
        Despite the significant accomplishments, using PLMs remains challenging and error-prone,
irrespective of their size. Firstly, the absence of a grounding mechanism complicates the
fact-checking of answers, particularly for tasks with an extractive nature, which are prone
to hallucination risks. Plus, many business automation workflows demand a high level of
accuracy and thus often incorporate human-in-the-loop interactions for auditing and correcting
predictions. This process necessitates knowledge about the precise location of the extracted
mentions in the text. Besides, disambiguation of the actual terms requires extra domain-specific
knowledge (such as soft skills [
        <xref ref-type="bibr" rid="ref15">15</xref>
        ]), making these processes as tedious and error-prone as their
predecessor equivalents (i.e., knowledge-driven applications using ESCO [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] and O*NET [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ]).
      </p>
      <p>To tackle the aforementioned issues, in this paper, we proposed a framework with OL to
extend and maintain ESCO. The primary contribution of this paper is the development and
implementation of a system capable of processing online job postings to extract skills, occupation
entities, and their corresponding relationships. Furthermore, the system can identify “new
entities" that are not yet included in ESCO, flagging them for further examination by a knowledge
or ontology engineer. Our methodology addresses multiple core aspects of OL to answer the
following research questions:
• RQ1: How efective is the proposed system in automated skill mention extraction from
online job postings?
• RQ2: How efectively is the proposed system classifying non-taxonomic relations between
skill and occupation types?
• RQ3: Is the proposed system capable of finding existing and/ or new entities that can
extend ESCO?</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Works</title>
      <p>
        OL addresses the challenges of knowledge acquisition and representation across various domains
[
        <xref ref-type="bibr" rid="ref16 ref7">16, 7</xref>
        ]. OL can be subdivided into several sub-processes, including the automatic identification
and extraction of terms, types, relations, and axioms from text. The study by [
        <xref ref-type="bibr" rid="ref17">17</xref>
        ] introduces
LLMs for OL using prompt-based learning. This approach leverages PLMs and cloze-style
language prompts to achieve promising results in various NLP tasks, such as sentiment
classification, knowledge probing, and natural language inference [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ].
      </p>
      <p>
        In their evaluation, [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ] assessed two open-source LLMs (Vicuna-13B and
Alpaca-LoRA13B with in-context learning) and a sentence transformer model (SBERT T5-XXL) using the
benchmark Text2KGBench [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. The results indicate high ontological conformance for both
Wikidata-TekGen and DBpedia-WebNLG corpora. However, these of-the-shelf LLMs performed
poorly on fact extraction, which the authors attribute to a lack of fine-tuning [
        <xref ref-type="bibr" rid="ref14">14</xref>
        ]. To bridge
the gap between semantic labelling tasks and text generation models for NER, [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] proposed
GPT-NER. This method transforms the NER task into a text-generation task and includes a
self-verification strategy to mitigate the excessive confidence of LLMs [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ]. The results show
performance comparable to fully supervised baselines based on BERT.
      </p>
      <p>
        Additionally, [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ] introduced LMDX, a methodology for using LLMs in information extraction,
particularly from visually rich documents. Their approach achieved a new state-of-the-art on
publicly available benchmarks such as CORD and VRDU [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]. Despite these notable findings, the
authors primarily focused on extracting mentions from text while grounding their predictions.
      </p>
      <p>
        In the labor market, many researchers leverage occupation ontologies to extract relevant
information from job posts. The work in [
        <xref ref-type="bibr" rid="ref21">21</xref>
        ] adapts a language model from ESCO to extract and
classify skill requirements from German-speaking job descriptions. The work in [
        <xref ref-type="bibr" rid="ref22">22</xref>
        ] detects
skills that are literally or implicitly mentioned in job ads and links them to ESCO. Besides, [
        <xref ref-type="bibr" rid="ref23">23</xref>
        ]
ifne-tuned the Llama model for extracting skills from job advertisements and user profiles. The
authors of [
        <xref ref-type="bibr" rid="ref24">24</xref>
        ] investigate the zero-shot approach for extracting skills based on ESCO. Despite
the aforementioned work, the work in [25] pre-trains a skill-aware language model usable for
domain-specific downstream tasks, such as job classification or skill extraction.
      </p>
      <p>
        There have been several works addressing OL in the labor market domain. The work by
[26], proposes NEO, a framework using approximately 2 million online job vacancies for the
enrichment of ESCO occupations. NEO identified 49 novel occupations of which 43 were
validated by an expert panel [26]. Furthermore, [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] proposed OntoJob, a cost-eficient unsupervised
framework that identifies and extracts knowledge, skill, ability, and competence mentions and
their corresponding relations using the C-value method and smoothed point-wise mutual
information (SPMI). [27] investigate the use of large language models for skill extraction leveraging
in-context learning to test two diferent prompting strategies. While there are parallels to the
work presented by us, [27] does not focus on knowledge discovery and relation classification.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Methodology</title>
      <p>To ensure the labor market ontology remains current and accurate, we propose a three-layered
framework: the knowledge layer, the back-end layer, and the front-end layer. The knowledge
ESCO</p>
      <p>Initialize</p>
      <p>Current
Knowledge</p>
      <p>Graph
Preprocessing Stage</p>
      <p>Extraction Stage</p>
      <p>Chunking
Prompt Construction</p>
      <p>Mention Extractor</p>
      <p>Mention
NO Mention YES</p>
      <p>Typing?</p>
      <p>Entity</p>
      <p>Linking
Colect
Metadata</p>
      <p>NO Familiar YES</p>
      <p>entity?
Propose new</p>
      <p>entity
Metadata</p>
      <p>New Entity</p>
      <sec id="sec-3-1">
        <title>Knowledge</title>
      </sec>
      <sec id="sec-3-2">
        <title>Layer</title>
      </sec>
      <sec id="sec-3-3">
        <title>Back-end</title>
      </sec>
      <sec id="sec-3-4">
        <title>Layer</title>
        <p>LaboDuraMtaarket</p>
      </sec>
      <sec id="sec-3-5">
        <title>Front-end</title>
      </sec>
      <sec id="sec-3-6">
        <title>Layer</title>
        <p>User Interface</p>
        <p>Post Processing Stage</p>
        <p>Relation
Classification</p>
        <p>Relation
NO New  YES</p>
        <p>Relation?
Colect
Metadata</p>
        <p>Propose new</p>
        <p>relation
Metadata</p>
        <p>New Relationship
layer houses the existing knowledge graph, initially based on established labor market graphs
such as ESCO. The back-end layer processes online job postings over specified intervals (e.g.,
weekly or monthly). This layer recommends new labor market mentions and entities, including
occupations and skills. The front-end layer serves as the user interface (UI), enabling a
humanin-the-loop mechanism via human annotators for updating the knowledge graph. Figure 1
illustrates each layer in updating the labor market’s knowledge graph. Our back-end architecture
is divided into three stages: (i) preprocessing, (ii) extraction, and (iii) postprocessing.</p>
        <sec id="sec-3-6-1">
          <title>3.1. The preprocessing stage</title>
          <p>
            To address input size limitation in LLMs, we split incoming data into manageable segments,
allowing the models to better process and understand context. Given document lengths and
PLM context window limitations, each job description text is divided into chunks. Then, The
prompt construction is used for formatting language model’s inputs and specifying how to
generate the output. Carefully designing prompts plays an essential role in specifying exactly
what sort of task and output format is expected. In our problem, for the prompt generation
stage, we largely adhere to the recommendations by [
            <xref ref-type="bibr" rid="ref20">20</xref>
            ]. We take the full set of documents
and apply a prompt template to format the document, beginning with “Document:". Next, we
append the task description and schema representation containing the entities to be extracted.
          </p>
          <p>
            The task description includes hard-coded instructions to guide the PLM in formatting the
output according to the schema. We provide the PLM with the following instruction: “Extract
[ENTITY TYPE] from the following document and format the output as a JSON with the following
structure: [SCHEMA]." In line with [
            <xref ref-type="bibr" rid="ref20">20</xref>
            ], the schema representation is a structured JSON object,
where the keys are the entity types to be extracted, and the values indicate their occurrence
(e.g., “" for a single instance and [] for multiple instances). For example, “occupation": “", “skill":
[] instructs the PLM to extract a single mention of the entity type “occupation" and multiple
mentions of the entity type “skill." In this context, a prompt template is a predefined format
used to structure input for the PLM, and the schema is a blueprint describing the structure and
organization of the data elements that need to be extracted.
          </p>
        </sec>
        <sec id="sec-3-6-2">
          <title>3.2. The extraction stage</title>
          <p>Once the input documents are processed in the preprocessing stage, we can begin extracting
the labor market mentions specified by the ESCO. Specifically, we focus on identifying the
ESCO skill entities, which involves locating skill mentions within the documents. However, our
method can be extended with other mention types like occupations or other entity types that
might occur in job postings.</p>
          <p>We define the mention extraction task to be two-fold, namely: (i) identification and extraction
of (sub-)strings in a given job posting, and (ii) indicating the term type of the given mention (i.e.,
is it a skill or occupation mention). In short, given a set of ESCO types  = {, },
we aim to find the mentions  of the types  from the total set of job posting documents .</p>
          <p>
            Similar to the design by [
            <xref ref-type="bibr" rid="ref20">20</xref>
            ], we use a decoder-only model for the extraction. A decoder-only
model predicts the next word to generate based on the previous words in the sequence. We use
the prompt instruction together with the desired output schema to instruct the PLM on what
entity mentions we want it to extract. We instruction-tune task-specific parameters in addition
to the pre-trained Mistral parameters on a data mixture containing a variety of (document,
schema, extraction) tuples [
            <xref ref-type="bibr" rid="ref20">28, 20</xref>
            ]. In contrast to the work by [
            <xref ref-type="bibr" rid="ref20">20</xref>
            ], we have a fixed schema
during our instruction tuning phase (i.e., we only train on a domain-specific data mixture).
Furthermore, all our documents are online job postings.
          </p>
          <p>Since we use a decoder-only model for “completing" the “output" with the correct mentions,
and types found in the actual job postings, extraction and typing happen at the same time.
However, we do want to separate the two diferent tasks since it is quite possible for the model
to extract mentions that are suitable for extraction but are put into the wrong category, which
leads to a mention typing mistake.</p>
          <p>Following the earlier instructions on the prompt construction and consequently, the task
description and the schema representation, the prompt for the model for the mention extraction
and entity linking tasks is as follows:
“&lt;s&gt;[INST] Extract {types} from the following document and format the output as a json with
the following structure: {format}
Document: {document} [/INST]
{output}&lt;/s&gt;"</p>
          <p>Then, we collect the mentions that are not being passed from the previous step to report
them to the human annotators as metadata that helps them annotate. Afterwards, given the
set of extracted mentions ℳ, and the set of ESCO concepts , we aim to map the extracted
mentions ℳ to the closest entity in the ESCO taxonomy. To achieve this, we define entity
linking to provide a many-to-one mapping for all mentions  ∈ ℳ to their corresponding
entity  ∈  as  () : ℳ ↦→ . To create the mapping  (.), we propose leveraging a retriever,
following the retrieval augmented generation design first proposed by [ 29], where for each
extracted mention, the top five closest entities are chosen. Given the extracted mention, we will
retrieve the approximate nearest ESCO entities using the retriever. Note that the index will be
entity-specific, meaning that the dense vector representations for ESCO skill entities are in a
diferent index than those for ESCO occupation entities. To map extracted skill mentions to
ESCO skills, we make use of both the “PreferedLabel" and the “AlternativeLabels" provided by
ESCO for each skill to make it easier for the retriever to retrieve the right skills, if they exist
in the taxonomy. Next, we leverage the results from the retriever and combine them with
instruction tuning [30]. We use the following prompt for the familiar entity check task:
“&lt;s&gt;[INST] Given a skill and options, select the best option that is a semantically exact
synonym for the skill. If none of the options is a semantically exact synonym, select ’No
Match’.</p>
          <p>Skill: {skill}
Options: {options}
Simply answer with the correct option with no explanation. [/INST]
{output}&lt;/s&gt;"</p>
          <p>In contrast to the work by [31], we also consider the “No Match" option to indicate that none
of the ESCO entities are a good match for the given mention.</p>
          <p>Essentially, the PLM is tasked with flagging the entity as “undiscovered" or mapping it
to one of the given ESCO entities provided by the retriever. The retriever acts as a filter
to limit the solution space of the matching to the most likely candidates, thus reducing the
| − |-classification to a 6-class classification problem instead.</p>
          <p>After classifying the found mentions in the entity linking, we use mentions that were marked
as “undiscovered" and occur frequently, to propose new entities to human annotators. When
the frequency of a mention exceeds a set threshold, we propose it as a new entity to the human
annotators, who can decide whether to add it as a new entity, add it as a synonym to an existing
entity, or not add it to the taxonomy at all if it is irrelevant.</p>
        </sec>
        <sec id="sec-3-6-3">
          <title>3.3. The postprocessing stage</title>
          <p>Conceptualization of the identified and extracted mentions is fourfold, namely; i) mention
extraction - identifying and extracting relevant terms, ii) entity linking - mapping the identified
and extracted mentions to their corresponding entities, iii) relationship extraction - identification
and extraction of the relations between the identified and extracted mentions, and lastly iv)
relationship classification - to map the identified and extracted relationships to their
domainspecific equivalent. We will primarily focus our attention on the non-taxonomical relations,
in particular: i) relationship(s) between mentions  ∈ ℳ, and entities  ∈  such that for
 ∈ , where  is the total set of relation types in our knowledge graph, we look for the triplet
relation  = (, , ) ∈ ℛ, and ii) relations between entities such that (, ,  ) ∈ ℛ, where
 ̸=  . We will refer to relations i) as classifying whether a mention is an alternative label
or synonym - for the entity in question, whereas we refer to ii) as finding out whether a skill
entity is essential, optional, or unrelated to an occupation entity.</p>
          <p>Since our research focuses on a subset of the entities in ESCO, we will also solely focus on
the possible links between these entities. As such, we will mainly look at the “IsOptionalFor",
and “IsEssentialFor" labels. In a similar fashion to the work by [31], we will construct a dataset
using the relations found in ESCO. Given that there are only three potential options, namely
 = { ,  , }, we opt for a similar approach to the
entity linking discussed earlier, but without a retriever (since there is no need to reduce the
number of classes, as was the case with the entity linking task). As such, we will task the PLM
to select the relation between a given skill and occupation entity from the set of relations .
We use the following prompt for the model for the relation classification task:
“&lt;s&gt;[INST] Given a skill and an occupation, tell me how important the skill is for the
occupation choosing from the following three options: essential, optional or not important.
Skill: {skill}
Occupation: {occupation}
Simply answer with the correct option with no explanation. [/INST]
{output}&lt;/s&gt;"</p>
          <p>Instruction training of the Mistral 7B model was done by leveraging the “AlternativeLabel"
and “PreferredLabel" data from ESCO in the generation of a train and test set. We used a true
“AlternativeLabel" related to the actual PreferredLabel as a positive example and randomly
sampled  − 1 non-related “AlternativeLabels" for negative examples. We would then use
this dataset and the instruction to train task-specific parameters in addition to the pre-trained
Mistral parameters for the entity linking task [28]. For more information on the evaluation and
implementation details, we refer to Section 4.</p>
          <p>In the current implementation, we only take into consideration two relations, namely: i)
“isOptionalSkillFor", and ii) “isEssentialSkillFor". Therefore, the check is relatively
straightforward: we see if the incoming entity pair (, ) are related via either i) or ii), or we
deem that the skill entity is not important to the occupation at all. If the relationship between
the (, ) pair did not exist, we propose a new relation with the given prediction.</p>
          <p>Similar to how we suggest new entities to human annotators, we also identify and propose
new relations between diferent entities in the taxonomy. By classifying relations between skill
entities and the occupation of the posting, we analyze currently non-existent relations. If a
frequently occurring new relation is found, we propose it to human annotators, who can then
decide whether to add it to the taxonomy.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Experiments</title>
      <p>This paper aims to leverage ESCO to cost-eficiently optimize and fine-tune PLMs for i) mention
extraction and term typing, ii) entity linking and knowledge discovery, and iii) relationship
classification. These fine-tuned PLMs, in turn, will help in the construction and maintenance of
the ontology and taxonomy. As described, we propose an evaluation of the full system and the
individual PLMs performance on each of these three tasks.</p>
      <sec id="sec-4-1">
        <title>4.1. Dataset</title>
        <p>In order to answer our research questions, we propose the three following experiments.</p>
        <p>Dataset
# total
# occupations
# skills
# essential
# optional
# negative</p>
        <p>ME
635
29,837</p>
        <p>Train</p>
        <p>EL
40,409</p>
        <p>10,617</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Experiment 1: Mention Extraction</title>
        <p>In order to extract skill mentions, we make us of the SkillSpan benchmark dataset, which
was provided by [32]. In particular, we employ the publicly available HOUSE, and TECH
data annotations, which contain, respectively, 90 and 110 job postings. In addition to the
SkillSpan dataset, we also make use of approximately 495 manually annotated job postings.
This proprietary dataset contains 1,058 chunks, which in total comprise 24,760 skills.</p>
        <p>As the data from the SkillSpan dataset and the proprietary dataset used internally were in
the BIO-tagging format, it was necessary to transform this data into a list of raw skill mentions,
as seen in the texts. It was essential to obtain the full raw skill mentions, as this is the type of
output that our extractor will be trained to extract. As our objective is to extract both hard and
soft skills, we employ both the “skill" and “knowledge" mentions as annotated in the SkillSpan
dataset. Furthermore, while the original data was segmented on a sentence level, we utilize the
provided vacancy index to convert the sentences back to their original job posting format.</p>
        <p>Subsequently, the postings were divided into sections of 384 tokens to ensure that the complete
prompt, along with the document and the expected output, would always fit within the context
window of the model. To assess the performance of both the base and fine-tuned models, we
conducted evaluations using all 58 job postings obtained from the SkillSpan test dataset. To
ensure reproducibility, no additional job postings were incorporated into the evaluation set.
As the input, we provided the model with complete job postings. Subsequently, the generated
output, which consists of the extracted raw skill mentions, was utilized to determine the F1 score
of the model in categorizing each token in the vacancy text as either a skill (1) or a non-skill
(0) token. The F1 score for the test set is presented in Table 2. For the mention extraction, we
train a LoRA of the Mistral-7B model. The model is trained on the job posting chunks for four
epochs and a batch size of eight. The ADAM optimizer is quantized with an 8-bit precision,
with 5 * 10− 5 learning rate, 50 warmup steps, and a weight decay of 0.01.
4.2.1. Baselines
For the mention extraction, we will utilize the most efective model from the study by [ 27] as a
baseline. This decision is primarily motivated by the fact that both our study and the study by
[27] use the evaluation data provided by [32]. Furthermore, [27] considers an extracted entity
correct even if it only partially overlaps with the gold span from the annotation. This aligns with
our metric, thus facilitating comparison. Additionally, we will compare the results of our model
to those discussed in the blog by [33]. The Skills Extractor Library, developed by [33], employs
a NER model based on spaCy’s architecture. This model maps the extracted skills to existing
taxonomies using semantic similarity. Conducting this comparison will allow us to evaluate
the performance of our model relative to their established skill extraction framework. Our last
baseline for the mention extraction experiment will be the models developed in the SkillSpan
paper [32]. The two models, one for “knowledge" extraction and one for “skill" extraction,
are BERT-based token classification models. We combine the results of both the “knowledge"
extractor and the “skill" extractor and consider tokens as either not a skill token or a skill token
without making any distinction between “knowledge" or “skill" or the beginning tokens (B) and
inside (I) tokens.</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Experiment 2: Relation Classification</title>
        <p>In the second experiment, the objective is to classify relations between skill and occupation pairs.
To construct the datasets for this experiment, we make use of all known relations between skills
and occupations in the ESCO database. For each skill/occupation relation, up to six variations
are included, utilizing the available alternative skill labels in ESCO. This encompasses the
optional and essential relations. Furthermore, for each of the existing relations, five random
skill-occupation combinations are sampled from the taxonomy. The aforementioned random
combinations will serve as the not important relation samples. Table 1 provides an overview
of the dataset distribution. We evaluate the system by computing the F1-score on the test set
detailed in Table 1.</p>
        <p>The training setup is as follows: A LoRA of the Mistral-7B model was trained. The model
was trained for 1500 steps with a batch size of 32 and 20 warmup steps. The Adam optimizer,
quantized to 8 bits, was employed with a learning rate of 5 * 10− 5 and a weight decay of 0.01.</p>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. Experiment 3: Knowledge Discovery</title>
        <p>To create the dataset for the entity classification dataset, we again make use of the ESCO
alternative labels for the skills. For each skill, we use up to 4 alternative labels to generate data
points. Each data point consists of the alternative label that is treated as the raw skill mention,
the preferred label as the correct answer, and the top 5 closest skills from the retriever. With
this information, we can fill in the prompt and expected output to train the model. For 40% of
the data points we show to the model, we will not include the correct skill as an option, and
instead, we show 5 incorrect options for which the target response will be “No Match" to allow
for the discovery of new skills or to discard mentions that should not be seen as a skill.</p>
        <p>To get insights into the performance of the proposed decoder-only model for i) linking
extracted skill mentions to ESCO skill entities, and ii) discovery of new potential ESCO skill
entities, we propose the following experimental setup. First, we will test the performance of the
model in the entity linking task by evaluating the F1 score on our test set.</p>
        <p>In addition, to know how the model performs on the discovery of new potential ESCO skill
entities, we manually annotate 1, 237 skill mentions extracted by our earlier stages in two
diferent stages. In the first stage, we manually annotate whether the skill mention matches one
of the five (i.e., assigning it the number of the suggestion) proposed suggestions by the retriever
or assign it a 6 if it matches none of the suggestions. Next, we filter out all the skill mentions
annotated with a 6, and check them for the following cases: i) the extracted mention itself does
not describe a skill (e.g., an occupation or organization name etc.), ii) the mention maps to one
(or more) existing ESCO skills but these options were not in the 5 suggestions, iii) the model
selected one or more good options for the list but the mention also includes additional skills that
are not in the top 5, iv) the mention is a proper skill mention, but ESCO does not contain any
suitable skills in the current taxonomy, and lastly, there is not enough context in the extracted
“mention" to judge (e.g., the mention just states “development").</p>
        <p>We train a LoRA of the Mistral-7B model. We train the model for 300 steps with a batch size
of 2 and 50 warmup steps. We use a quantized Adam 8-bit optimizer with a learning rate of
2.5 * 10− 5 and a weight decay of 0.01. Retrieval of the five “closest" skills, is done with dense
retrieval. For embedding the skills, we used MixedBread without any additional fine-tuning.
4.4.1. Baselines
To evaluate the performance of our mapper, we will compare its results to those of the mapper
introduced by [33]. The Skills Extractor Library developed by [33] utilizes MiniLM to encode
skills and map them to the closest ESCO entity. To assess the efectiveness of our approach, we
use the same dataset with the ESCO alternative labels but without “No Match" data points. We
employ the evaluation method from [33] and present the F1 score in Table 2 for comparative
analysis. It is worthless that in this case, the mapper from [33] solves a simpler problem
compared to ours since it does not consider “No Match".</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Results</title>
      <sec id="sec-5-1">
        <title>5.1. Experiment 1: Mention Extraction</title>
        <p>Table 2 demonstrates that the Mistral 7B model (the Base model) has an F1 score of 0.28 for
skill mention extraction across 58 job postings. Following the instruction tuning of a Low-Rank
Adaptation (LoRA) model on 635 manually annotated job postings (denoted as Base + ME),
evaluation on the same set of 58 job postings results in an F1 score of 0.54. This indicates that
instruction tuning on self-supervised labor market ontology data enhances performance by
approximately 0.26. We also conducted a comprehensive comparison by evaluating three
existing methods on the same dataset. Specifically, the model proposed by [ 27] yielded an F1
score of 0.46, while the baseline model discussed in [33] achieved one of 0.27. Notably, the
model developed by [32] outperformed our Base + ME model, achieving an F1 score of 0.80 on
the evaluation dataset. These results suggest that our model’s performance is comparable to,
but not in all cases superior to, other existing methodologies on this task.</p>
        <p>Model ↓ / Task →
[27]*
[33]*
[32]*</p>
        <p>Base
Base + ME
Base + RC
Base + EL</p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Experiment 2: Relation Classification</title>
        <p>Results from Table 2 show that the Base model scored a F1 score of 0.54. The instruction
finetuned LoRA model, the so-called “Base + RC", scored an F1 of 0.66. Results include incorporation
of negative examples as dictated in Table 1, and shufling the options into a random ordering.
In total, we see that instruction tuning leads to an approximate performance increase of 0.12.</p>
      </sec>
      <sec id="sec-5-3">
        <title>5.3. Experiment 3: Knowledge Discovery</title>
        <p>The third experiment was essentially twofold. Firstly, we report the results of evaluating the
parsed skill mentions from the job posting test set. Table 2 shows us that the “Base + EL" model
scored an F1 score of 0.67, while the Base model only scored an F1 score of 0.30. As such,
instruction tuning the Mistral Base leads to an increased performance of approximately 0.37.
Besides, we compared our method with the baseline model described in [33], which yielded an
F1 score of 0.57. In this case, “Base + EL" model showed a better result, despite the fact that
it solved a more complex problem, not only identifying the most fitting ESCO skills, but also
indicating potentially new ones.</p>
        <p>Results for the manual annotation of the mentions marked as “new entities" showed an F1
score of 0.41. In the cases where the extracted mention is not a valid skill mention or there is
not enough context, the model shows an F1 score of 0.42. Lastly, in the case where the extracted
mention can map to an existing ESCO skill but this ESCO skill was not included as one of the 5
options that the model could select from, the model obtains an F1 score of only 0.16.</p>
        <p>Manual annotation of the 1, 237 extracted skill mentions showed that 704 of those mentions
could not be linked to one of the five provided suggestions of the retriever. Careful examination
of the 704 skill mentions that could not be linked to the five provided suggestions showed that
94 had an existing ESCO entity, but the retriever failed to select the appropriate entity for the
suggestion. From the 704 total skill mentions 282 were manually annotated to be a potential
“New Skill". Additionally, 136 were not actual skill types, and 160 lacked the context to make a
valid prediction. In total, the model extracted a total of 253 skill mentions that were annotated
as a potential new skill to be reviewed by human annotators for addition to the ESCO taxonomy.
A few examples of these mentions are: “ReactJS", “AWS", and “Docker".</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Discussion</title>
      <p>Our paper explored three diferent research questions. To address RQ1, we instruction-tuned a
LoRA model to the Mistral Base on a total of 635 job postings. According to 2, the “Base + ME"
model outperforms the “Base" model in extracting the skill mentions by scoring an F1 of 0.54
compared to 0.28. Accordingly, although the “Base" model struggles with the extraction of the
skill mentions, decoder-only architectures can be instruction-tuned to extract and format skill
mentions from raw job posting texts. Plus, on the ME task, the "Base + ME" model outperforms
the best-performing model from [27] by 0.8. While it is dificult to credit this diference solely
to the proposed system (i.e., this would require more details on the actual diferences between
gpt-3.5 turbo and the Mistral Base model), we believe that it demonstrates the competitiveness
of our proposed system with the state-of-the-art. We still see that the BERT-based models from
[32] outperform our proposed method. However their proposed system has a major drawback
that it requires BIO-tagged data on the token level to train, which is labour-intensive to obtain
compared to just having to obtain a list of mentions in the text like we need for our approach.</p>
      <p>To answer RQ2, we performed the relation classification which determines if a skill entity
is “optional", “essential", or “not important" to be added ESCO. The results indicate that the
autoregressive model is capable of learning the relation classification task via self-supervised
instruction tuning from ESCO. However, we only trained the Base + RC model on 32.000
examples due to time constraints. Thus, there exist opportunities to improve the current
model’s performance by training with more examples.</p>
      <p>To the best of our knowledge, there is no other study that looks at the relation classification
between ESCO skill, and occupation entities using autoregressive models as proposed in this
work. However, the work by [31] can be regarded as very similar. [31] perform entity
classiifcation, and relation classification at once, therefore, we can’t use their F1 scores for direct
comparison. Having said that, the "Base + RC" models’ performance appears to be on par with
the F1 score of 0.51, outperforming by 0.15 on the slightly diferent task. [ 31] has a model
that predicts both the types of the subject and object and the predicate while constraining the
possible labels to a predefined set (i.e., choose between skill and occupation for the entity type).
On the other hand, the "Base + RC" model only predicts the predicate.</p>
      <p>Lastly, to answer RQ3, in the knowledge discovery experiment, we consider two diferent
experiments. Results from the first experiment help us figure out whether instruction tuning
the “Base" model with self-supervised data from ESCO would increase the performance on the
entity-linking task. The “Base + EL" model scores approximately 0.37 above the “Base" model,
demonstrating the efectivity of self-supervised instruction tuning using ESCO. Additionally,
having a more complex task, which includes, in addition to entity linking, the indication of not
familiar entities, the "Base + EL" model outperforms the method proposed in [33] by 0.1. This
demonstrates that our approach is highly competitive with other methods in entity linking.</p>
      <p>The second experiment grants us insight into the ability of the decoder-only model to
augment and enrich the ESCO with skill mention suggestions for human annotators. The “Base
+ EL" model suggested 1, 054 skill mentions with no matching ESCO entity, whereas manual
annotation by 6 human domain experts revealed only 704. However, there was no further
manual annotation of the 1, 054 mentions predicted to have “No Match" and the quality of
the suggestions provided by the retriever. We believe that similarly to the 704 mentions that
received annotation, there will be a proportion where the provided suggestions by the retriever
were wrong (i.e., 94 out of the 704 for which we provided manual annotation had existing ESCO
skill entities that were not suggested as an option by the retriever).</p>
      <p>Overall, the “Base + EL" models are capable of selecting mentions that have no match in
ESCO. However, there is a need for post-processing steps to filter out “false positives" inter
alia; the extracted mention is not a “skill", lacks the full context to make a prediction, and was
wrongly classified as “No Match" due to missing suggestion. For this study, post-processing was
done manually, leading to flagging 253 extracted skill mentions as a potential addition to ESCO.</p>
    </sec>
    <sec id="sec-7">
      <title>7. Implications and Limitations</title>
      <p>The results from this paper demonstrate the efectiveness and adaptability of using LoRA and
decoder-only architectures for ontology learning in the labor market setting. Results indicate
that the generation of self-supervised datasets, combined with instruction tuning, leads to
impressive performance gains on skill mention extraction, relation classification (i.e., on skill
and occupation entity pairs), and lastly, discovery of skill mentions that could potentially extend
the labor market ontology/ taxonomy. We believe the unification of LLMs and ontologies can aid
in the lacking abilities of knowledge persistence in LLMs. Since the labor market is a constantly
changing environment, editing knowledge without re-training the whole LLM is of utmost
importance. The proposed system provides an intuitive way to enrich and maintain ontologies
(not just labor market specific), while at the same time leveraging the knowledge of the ontology
to keep the models up-to-date by creating self-supervised training sets.</p>
      <p>Furthermore, we believe that our results demonstrated the potential strength of using a
proposed system to augment and enrich existing labor market ontologies and/ or taxonomies
(i.e., ESCO, the O*NET, etc.). In particular, our results show the pivotal role of the retriever in
easing the construction of self-supervised data that the decoder-only model easily leverages via
instruction tuning. Additionally, our models demonstrate utility in alleviating the time- and
resource constraints in human annotation by training “smaller" language models to assist (i.e.,
models that fit on a single GPU).</p>
      <p>The current study has some limitations to be considered. Firstly, for entity linking, we tried
leveraging the skill attribute “AlternativeLabel" as provided by ESCO. However, we did so under
the assumption that the listed alternative labels are in a way synonyms to the “PreferredLabel".
This assumption does not necessarily hold in reality. Secondly, the current study did not
experiment with hyperparameter tuning during model training and evaluation. There is considerable
room for improvement of the models via heyperparameter tuning.</p>
      <p>
        Thirdly, another limitation in the entity-linking experiment is that we do not consider the full
context of the job posting when linking the entity. For example, the extracted mention
“engineering", gets five valid suggestions, namely; “software engineering", “packaging engineering",
and “power engineering". Since there is no context, none of the five options is more valid than
the other. Finally, our current implementation does not incorporate any knowledge-grounding
methodologies. During the study, we experimented with the incorporation of index values to
ground the extractions by similarly checking the index to the work by [
        <xref ref-type="bibr" rid="ref20">20</xref>
        ]. However, Mistral
7B seemed to have trouble with the provided index values.
      </p>
    </sec>
    <sec id="sec-8">
      <title>8. Conclusion</title>
      <p>We introduce an OL system to upgrade the existing labor market ontologies. We analyzed the
job postings to extract new entities such as occupations and skills. We propose a framework to
recommend entity linking between skills and occupations. To evaluate the performance, we
designed multiple experiments to address our research questions about the performance of skill
extraction, non-taxonomical relationship retrieval, and knowledge discovery.</p>
      <p>As future work, we add “post-processing" filters to better distinguish the diferent types
of non-matches would be valuable. This potentially saves the human annotator from sifting
through mentions that are amongst other things; not skill types, too vague, and/ or present
in ESCO. Furthermore, there is considerable room for improvement on the hyperparameter
settings used in training the models in this paper, we consider this one of the easiest avenues for
improvement of the results. Lastly, we would be very interested in testing out the current system
on a variety of diferent types that are currently not part of ESCO. For example, extracting wage
information, educational requirements, benefits, requirements about work experience, etc.
[25] C. Fang, C. Qin, Q. Zhang, K. Yao, J. Zhang, H. Zhu, F. Zhuang, H. Xiong, Recruitpro: A
pretrained language model with skill-aware prompt learning for intelligent recruitment,
in: Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data
Mining, KDD ’23, ACM, 2023.
[26] A. Giabelli, L. Malandri, F. Mercorio, M. Mezzanzanica, A. Seveso, Neo: A tool for taxonomy
enrichment with new emerging occupations, in: International Semantic Web Conference,
Springer, 2020, pp. 568–584.
[27] K. C. Nguyen, M. Zhang, S. Montariol, A. Bosselut, Rethinking skill extraction in the job
market domain using large language models, arXiv preprint arXiv:2402.03832 (2024).
[28] E. J. Hu, Y. Shen, P. Wallis, Z. Allen-Zhu, Y. Li, S. Wang, L. Wang, W. Chen, Lora: Low-rank
adaptation of large language models, arXiv preprint arXiv:2106.09685 (2021).
[29] P. Lewis, E. Perez, A. Piktus, F. Petroni, V. Karpukhin, N. Goyal, H. Küttler, M. Lewis, W.-t.</p>
      <p>Yih, T. Rocktäschel, et al., Retrieval-augmented generation for knowledge-intensive nlp
tasks, Advances in Neural Information Processing Systems 33 (2020) 9459–9474.
[30] J. Wei, M. Bosma, V. Y. Zhao, K. Guu, A. W. Yu, B. Lester, N. Du, A. M. Dai, Q. V. Le,</p>
      <p>Finetuned language models are zero-shot learners, arXiv preprint arXiv:2109.01652 (2021).
[31] J. Vrolijk, D. Graus, Enhancing plm performance on labour market tasks via
instructionbased finetuning and prompt-tuning with rules, arXiv preprint arXiv:2308.16770 (2023).
[32] M. Zhang, K. N. Jensen, S. D. Sonniks, B. Plank, Skillspan: Hard and soft skill extraction
from english job postings, arXiv preprint arXiv:2204.12811 (2022).
[33] E. Gallagher, I. Kerle, C. Sleeman, J. Vines, The skills extractor library, 2023. Accessed:
2024-06-06.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>J. D.</given-names>
            <surname>Smedt</surname>
          </string-name>
          , M. le
          <string-name>
            <surname>Vrang</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Papantoniou</surname>
          </string-name>
          , Esco:
          <article-title>Towards a semantic web for the european labor market</article-title>
          ,
          <source>in: LDOW@WWW</source>
          ,
          <year>2015</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>E.</given-names>
            <surname>Commission</surname>
          </string-name>
          ,
          <article-title>Esco handbook european skills, competences, qualifications and occupations, Publications Ofice of the EU (</article-title>
          <year>2019</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>National</given-names>
            <surname>Center for O*NET Development.</surname>
          </string-name>
          ,
          <string-name>
            <surname>O*</surname>
          </string-name>
          <article-title>net online</article-title>
          .,
          <year>2024</year>
          . Online; accessed 4-
          <fpage>June2024</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>J.</given-names>
            <surname>Djumalieva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Sleeman</surname>
          </string-name>
          ,
          <article-title>An open and data-driven taxonomy of skills extracted from online job adverts, in: Developing skills in a changing world of work</article-title>
          , Rainer Hampp Verlag,
          <year>2018</year>
          , pp.
          <fpage>425</fpage>
          -
          <lpage>454</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>S.</given-names>
            <surname>Pan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <article-title>Unifying large language models and knowledge graphs: A roadmap, IEEE Transactions on Knowledge and Data Engineering (</article-title>
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Sibarani</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Scerri</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Morales</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Auer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Collarana</surname>
          </string-name>
          ,
          <article-title>Ontology-guided job market demand analysis: a cross-sectional study for the data science field</article-title>
          ,
          <source>in: Proceedings of the 13th international conference on semantic systems</source>
          ,
          <year>2017</year>
          , pp.
          <fpage>25</fpage>
          -
          <lpage>32</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>P.</given-names>
            <surname>Buitelaar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Cimiano</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Magnini</surname>
          </string-name>
          ,
          <article-title>Ontology learning from text: methods, evaluation and applications</article-title>
          , volume
          <volume>123</volume>
          , IOS press,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>J.</given-names>
            <surname>Vrolijk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. T.</given-names>
            <surname>Mol</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Weber</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Tavakoli</surname>
          </string-name>
          , G. Kismihók,
          <string-name>
            <given-names>M.</given-names>
            <surname>Pelucchi</surname>
          </string-name>
          ,
          <article-title>Ontojob: Automated ontology learning from labor market data</article-title>
          ,
          <source>in: 2022 IEEE 16th International Conference on Semantic Computing (ICSC)</source>
          , IEEE,
          <year>2022</year>
          , pp.
          <fpage>195</fpage>
          -
          <lpage>200</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>K.</given-names>
            <surname>Frantzi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Ananiadou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Mima</surname>
          </string-name>
          ,
          <article-title>Automatic recognition of multi-word terms:. the c-value/nc-value method</article-title>
          ,
          <source>International journal on digital libraries 3</source>
          (
          <year>2000</year>
          )
          <fpage>115</fpage>
          -
          <lpage>130</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>S.</given-names>
            <surname>Roller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kiela</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Nickel</surname>
          </string-name>
          ,
          <article-title>Hearst patterns revisited: Automatic hypernym detection from large text corpora</article-title>
          , arXiv preprint arXiv:
          <year>1806</year>
          .
          <volume>03191</volume>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>H.</given-names>
            <surname>Mousavi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kerr</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Iseli</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zaniolo</surname>
          </string-name>
          ,
          <article-title>Harvesting domain specific ontologies from text</article-title>
          ,
          <source>in: 2014 IEEE International Conference on Semantic Computing, IEEE</source>
          ,
          <year>2014</year>
          , pp.
          <fpage>211</fpage>
          -
          <lpage>218</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>T.</given-names>
            <surname>Brown</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Mann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ryder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Subbiah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. D.</given-names>
            <surname>Kaplan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Dhariwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Neelakantan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Shyam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Sastry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Askell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Herbert-Voss</surname>
          </string-name>
          , G. Krueger,
          <string-name>
            <given-names>T.</given-names>
            <surname>Henighan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Child</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Ramesh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Ziegler</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Winter</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Hesse</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chen</surname>
          </string-name>
          , E. Sigler,
          <string-name>
            <given-names>M.</given-names>
            <surname>Litwin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Gray</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Chess</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Clark</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Berner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>McCandlish</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Radford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Sutskever</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Amodei</surname>
          </string-name>
          ,
          <article-title>Language models are few-shot learners</article-title>
          , in: H.
          <string-name>
            <surname>Larochelle</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Ranzato</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Hadsell</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Balcan</surname>
          </string-name>
          , H. Lin (Eds.),
          <source>Advances in Neural Information Processing Systems</source>
          , volume
          <volume>33</volume>
          ,
          <string-name>
            <surname>Curran</surname>
            <given-names>Associates</given-names>
          </string-name>
          , Inc.,
          <year>2020</year>
          , pp.
          <fpage>1877</fpage>
          -
          <lpage>1901</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>T.</given-names>
            <surname>Schick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Schütze</surname>
          </string-name>
          ,
          <article-title>It's not just size that matters: Small language models are also few-shot learners</article-title>
          , in: K.
          <string-name>
            <surname>Toutanova</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Rumshisky</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Zettlemoyer</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Hakkani-Tur</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <string-name>
            <surname>Beltagy</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Bethard</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Cotterell</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Chakraborty</surname>
          </string-name>
          , Y. Zhou (Eds.),
          <source>Proceedings of the</source>
          <year>2021</year>
          <article-title>Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Association for Computational Linguistics</article-title>
          , Online,
          <year>2021</year>
          , pp.
          <fpage>2339</fpage>
          -
          <lpage>2352</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>N.</given-names>
            <surname>Mihindukulasooriya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Tiwari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. F.</given-names>
            <surname>Enguix</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lata</surname>
          </string-name>
          ,
          <article-title>Text2kgbench: A benchmark for ontology-driven knowledge graph generation from text</article-title>
          , in: International Semantic Web Conference, Springer,
          <year>2023</year>
          , pp.
          <fpage>247</fpage>
          -
          <lpage>265</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>D.</given-names>
            <surname>Beauchemin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Laumonier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y. L.</given-names>
            <surname>Ster</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yassine</surname>
          </string-name>
          ,
          <article-title>"fijo": a french insurance soft skill detection dataset</article-title>
          ,
          <year>2022</year>
          . arXiv:
          <volume>2204</volume>
          .
          <fpage>05208</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>A.</given-names>
            <surname>Konys</surname>
          </string-name>
          ,
          <article-title>Knowledge repository of ontology learning tools from text</article-title>
          ,
          <source>Procedia Computer Science</source>
          <volume>159</volume>
          (
          <year>2019</year>
          )
          <fpage>1614</fpage>
          -
          <lpage>1628</lpage>
          .
          <source>Knowledge-Based and Intelligent Information &amp; Engineering Systems: Proceedings of the 23rd International Conference KES2019.</source>
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>H.</given-names>
            <surname>Babaei Giglou</surname>
          </string-name>
          ,
          <string-name>
            <surname>J. D'Souza</surname>
            ,
            <given-names>S. Auer,</given-names>
          </string-name>
          <article-title>Llms4ol: Large language models for ontology learning</article-title>
          , in: International Semantic Web Conference, Springer,
          <year>2023</year>
          , pp.
          <fpage>408</fpage>
          -
          <lpage>427</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>N.</given-names>
            <surname>Ding</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chen</surname>
          </string-name>
          , X. Han,
          <string-name>
            <surname>G</surname>
          </string-name>
          . Xu,
          <string-name>
            <given-names>P.</given-names>
            <surname>Xie</surname>
          </string-name>
          , H.-T. Zheng,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          , H.-G. Kim,
          <article-title>Promptlearning for fine-grained entity typing</article-title>
          ,
          <source>arXiv preprint arXiv:2108.10604</source>
          (
          <year>2021</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>S.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Ouyang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <surname>G.</surname>
          </string-name>
          <article-title>Wang, Gpt-ner: Named entity recognition via large language models</article-title>
          ,
          <source>arXiv preprint arXiv:2304.10428</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>V.</given-names>
            <surname>Perot</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Kang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Luisier</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. S.</given-names>
            <surname>Boppana</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Mu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , N. Hua,
          <article-title>Lmdx: Language model-based document information extraction and localization</article-title>
          ,
          <source>arXiv preprint arXiv:2309.10952</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <surname>A</surname>
          </string-name>
          .
          <article-title>-s.</article-title>
          <string-name>
            <surname>Gnehm</surname>
            , E. Bühlmann,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Buchs</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Clematide</surname>
          </string-name>
          ,
          <article-title>Fine-grained extraction and classification of skill requirements in German-speaking job ads</article-title>
          , in: D.
          <string-name>
            <surname>Bamman</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Hovy</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Jurgens</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Keith</surname>
            ,
            <given-names>B. O</given-names>
          </string-name>
          <string-name>
            <surname>'Connor</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          Volkova (Eds.),
          <source>Proceedings of the Fifth Workshop on Natural Language Processing and Computational Social Science (NLP+CSS)</source>
          ,
          <article-title>Association for Computational Linguistics, Abu Dhabi</article-title>
          ,
          <string-name>
            <surname>UAE</surname>
          </string-name>
          ,
          <year>2022</year>
          , pp.
          <fpage>14</fpage>
          -
          <lpage>24</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <year>2022</year>
          . nlpcss-
          <volume>1</volume>
          .2.
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <surname>J.-J. Decorte</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Verlinden</surname>
            ,
            <given-names>J. V.</given-names>
          </string-name>
          <string-name>
            <surname>Hautte</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Deleu</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Develder</surname>
          </string-name>
          , T. Demeester,
          <article-title>Extreme multilabel skill extraction training using large language models</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2307</volume>
          .
          <fpage>10778</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>N.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Kang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T. D.</given-names>
            <surname>Bie</surname>
          </string-name>
          ,
          <article-title>Skillgpt: a restful api service for skill extraction and standardization using a large language model</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2304</volume>
          .
          <fpage>11060</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>B.</given-names>
            <surname>Clavié</surname>
          </string-name>
          , G. Soulié,
          <article-title>Large language models as batteries-included zero-shot esco skills matchers</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2307</volume>
          .
          <fpage>03539</fpage>
          .
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>