<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <journal-title-group>
        <journal-title>URL: https://arxiv.org/abs/</journal-title>
      </journal-title-group>
    </journal-meta>
    <article-meta>
      <article-id pub-id-type="doi">10.18653/v1/2022.acl-long.62</article-id>
      <title-group>
        <article-title>AlexU-NLP at TalentCLEF 2025: Curriculum-Driven Hybrid Retrieval for Multilingual Job Title Matching</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Rana Barakat</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Omar Mokhtar</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Marwan Torki</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Nagwa Elmakky</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Computer and Systems Engineering Department, Alexandria University</institution>
          ,
          <country country="EG">Egypt</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2409</year>
      </pub-date>
      <volume>10173</volume>
      <fpage>878</fpage>
      <lpage>891</lpage>
      <abstract>
        <p>This paper describes our approach for TalentCLEF 2025 Task A, focusing on multilingual and cross-lingual job title matching. The core challenge lies in the inherent brevity and ambiguity of job titles across diferent languages (English, Spanish, and German) and professional sectors. Our methodology employs a curriculum learning strategy to fine-tune an embedding model, gradually exposing it to more complex data involving job titles and their descriptions. We further enhance retrieval performance through a hybrid system combining semantic search with BM25 keyword matching, followed by a multilingual cross-encoder reranker. Experimental results on the validation set demonstrate the efectiveness of our phased training approach and hybrid retrieval, achieving a top average mAP of 56% on the validation set and an average mAP of 53% on the test set.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Embedding Models</kwd>
        <kwd>Hybrid Retrieval</kwd>
        <kwd>Curriculum Learning</kwd>
        <kwd>Large Language Models</kwd>
        <kwd>Human Resources</kwd>
        <kwd>TalentCLEF</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>The modern workplace has undergone a profound transformation in recent years, driven by technological
innovation, globalization, and shifting social dynamics. Technological advancements—particularly in
artificial intelligence and natural language processing—are reshaping how companies source, assess,
and manage human capital. At the same time, the globalization of the workforce, enabled by remote
hiring and digital collaboration tools, has introduced new complexities in matching candidates to job
roles across linguistic and cultural boundaries. These developments require intelligent systems capable
of handling large-scale, multilingual data while maintaining the semantic integrity of role descriptions
and candidate profiles.</p>
      <p>A central challenge in this space is the variability and ambiguity of job titles. Job titles are often brief,
under-specified, and highly context-dependent.</p>
      <p>Furthermore, diferent organizations frequently use distinct terms to describe similar roles. For
instance, the positions of “Software Engineer,” “Backend Developer,” and “Platform Engineer” may
share significant overlap in responsibilities, yet difer in naming conventions based on organizational or
regional preferences. This terminological inconsistency becomes even more pronounced in multilingual
contexts, where translation, cultural nuance, and domain-specific jargon further complicate the task of
semantic alignment.</p>
      <p>
        Task A of TalentCLEF 2025 [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] addresses this problem by focusing on multilingual and cross-lingual
job title matching across English, Spanish, and German. The objective is to retrieve and rank relevant job
titles for a given query title, leveraging both linguistic and contextual understanding. Efective solutions
to this task must reconcile lexical variation, cross-language equivalence, and domain specificity while
remaining computationally eficient and scalable.
      </p>
      <p>In this notebook, we present AlexU-NLP’s approach to Task A, which combines curriculum learning,
hybrid retrieval, and re-ranking to address the multifaceted challenges of job title normalization
and retrieval. Our methodology employs a curriculum-based fine-tuning strategy for a multilingual
embedding model, progressively introducing more complex data—from isolated job titles to rich
titledescription pairs—thereby enhancing the model’s capacity to learn meaningful semantic representations.
To maximize retrieval performance, we adopt a hybrid strategy that integrates dense semantic search
with BM25-based sparse retrieval, followed by a multilingual cross-encoder re-ranker.</p>
      <p>This system not only achieves strong performance on the TalentCLEF 2025 validation and test sets
but also demonstrates practical relevance in real-world Human Capital Management (HCM) scenarios.
By improving the robustness and adaptability of job title matching systems, our approach supports
more accurate talent identification and enhances the alignment between workforce capabilities and
organizational needs in a multilingual, cross-sector labor market.</p>
    </sec>
    <sec id="sec-2">
      <title>2. Related Work</title>
      <p>
        In the evolving landscape of job recommendation systems, recent research has focused on enhancing
the semantic understanding of job titles and descriptions to improve candidate-job matching. Zbib
et al.[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] introduced an unsupervised method that learns job title similarities by leveraging noisy skill
labels, demonstrating efectiveness in text ranking and job normalization tasks.
      </p>
      <p>
        Complementing this, Laosaengpha et al. [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] proposed a Job Description Aggregation Network (JDAN)
that derives job title representations directly from job descriptions, bypassing the need for explicit skill
extraction and achieving superior performance over traditional skill-based approaches. Addressing
multilingual challenges, Zhang et al.[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] developed ESCOXLM-R, a multilingual language model
pretrained on the ESCO taxonomy across 27 languages, which achieved state-of-the-art results on various
job-related tasks.
      </p>
      <p>
        Furthermore, Deniz et al. [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] combined unsupervised and contrastive learning techniques to create
a multilingual job title encoder, enhancing cross-lingual job recommendation capabilities. These
advancements collectively contribute to more accurate and inclusive job matching systems in a global
context.
      </p>
    </sec>
    <sec id="sec-3">
      <title>3. Dataset</title>
      <p>
        The corpus for Task A [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] comprises job titles in English, Spanish, and German, spanning various job
domains and professional sectors.
      </p>
      <p>• Training Data: Provided as 15,000 pairs of related job titles per language (English, Spanish,</p>
      <p>German).
• Validation Data: Structured into three distinct files per language: queries, corpus elements, and
qrels (query relevance assessments). This set contains 100 query job titles per language, each
with a list of related job titles from the corpus. A knowledge base of 2,500 unique job titles per
language serves as the corpus for retrieval tasks within the validation set.
• Test Data: A background set comprising 5,000 job titles. The evaluation is conducted on a subset
of the background set, that will be a gold standard corpus of 100 job titles in each language.</p>
    </sec>
    <sec id="sec-4">
      <title>4. Methodology</title>
      <sec id="sec-4-1">
        <title>4.1. Data Augmentation</title>
        <p>
          To enhance the contextual understanding of job titles during model fine-tuning and inference, we
implemented a two-pronged data augmentation strategy. First, for enriching the training dataset, we
used the European Skills, Competences, Qualifications and Occupations (ESCO) taxonomy [
          <xref ref-type="bibr" rid="ref7">7</xref>
          ]. We
utilized this resource to source authentic job descriptions corresponding to the titles in our training set.
The integration of these descriptions furnished the model with rich, real-world contextual information,
which is vital for efective fine-tuning. Second, to address the common challenge of missing descriptions
for corpus entries during the inference phase, we employed the Qwen3-14B large language model
(LLM) [
          <xref ref-type="bibr" rid="ref8">8</xref>
          ]. This model was tasked with generating synthetic yet contextually plausible job descriptions
for each title within the inference corpus using zero-shot prompting. This ensured that every entry in
our retrieval corpus, both for validation and testing, consisted of a title paired with a description, a step
whose impact is quantified in our results (Section 6.4).
        </p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Embedding Model and Fine-tuning Rationale</title>
        <p>
          We selected the multilingual-E5-large-instruct model [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ] as our core embedding backbone
due to its strong performance on multilingual tasks and its instruction-tuned architecture, which is
beneficial for understanding task-specific nuances. Efectively fine-tuning such models for our task
requires careful consideration, as simplistic fine-tuning strategies can present certain challenges. For
instance, training exclusively on brief job title pairs risks overfitting to lexical patterns, limiting the
model’s ability to generalize to semantically equivalent but lexically diverse titles. Conversely, directly
incorporating lengthy job descriptions from the start might lead the model to become overly reliant on
this rich contextual data, diminishing its focus on the job titles themselves and impacting performance
when such descriptions are absent or of variable quality. To navigate these pitfalls and foster a more
balanced and robust learning process, we adopted a curriculum learning strategy, detailed in the
subsequent section.
        </p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Curriculum Learning Implementation</title>
        <p>We implemented a curriculum learning approach that incrementally increases the complexity of the
training data. This allows the model to first establish a robust understanding of job title semantics before
integrating the richer contextual information from job descriptions. A key aspect of our methodology
is its emphasis on cross-lingual learning. For instance, for each English job title, our training data
included not only pairs with its relevant English title and description but also pairs with translations of
both the title and its description into German and Spanish. This systematic exposure to semantically
equivalent information across languages is designed to encourage the alignment of embeddings in the
multilingual space, thereby mitigating language-specific clustering. The curriculum was structured
into several stages. These are briefly described in the following subsections, while full details regarding
their specific configurations can be found in Section 5.4.</p>
        <sec id="sec-4-3-1">
          <title>4.3.1. Initial Stages</title>
          <p>The curriculum commenced with training on symmetric (job title, job title) pairs. These initial stages
prioritized monolingual data, with the model being trained on pairs within each language (en-en, es-es,
de-de) before progressing to cross-lingual title pairs (en-es, en-de). The primary objective of these early
stages was to preserve and refine the pretrained model’s intrinsic ability to align job titles based purely
on their semantics, forming a strong foundation for subsequent learning.</p>
        </sec>
        <sec id="sec-4-3-2">
          <title>4.3.2. Intermediate Stages</title>
          <p>Following the initial alignment, the curriculum gradually introduced asymmetric pairs of (job title, job
title + job description). This was done in a controlled manner, balancing these richer contextual pairs
with the title-only examples from the preceding stages. Similar to the early stages, this phase also began
with same-language pairs before incorporating cross-lingual pairs. This part of the curriculum acted
as a form of soft domain adaptation, injecting more extensive contextual cues to help disambiguate
job titles and discourage overfitting to short, potentially ambiguous title tokens, while simultaneously
reinforcing the symmetric retrieval structure learned earlier.</p>
        </sec>
        <sec id="sec-4-3-3">
          <title>4.3.3. Final Stages</title>
          <p>The concluding stages of the curriculum placed a strong emphasis on the asymmetric (job title, job
title + job description) format. This was intended to allow the model to fully adapt to the anticipated
real-world inference conditions, where job descriptions are expected to provide significant contextual
information.</p>
        </sec>
      </sec>
      <sec id="sec-4-4">
        <title>4.4. Hybrid Retrieval and Reranking</title>
        <p>Our retrieval architecture employs a two-stage process, consisting of an initial hybrid retrieval phase
followed by a neural reranking mechanism.</p>
        <p>In the first stage, we perform a hybrid search by combining signals from dense and sparse retrieval
methods.</p>
        <p>• For dense retrieval, query job titles and corpus entries (each comprising a job title and
its corresponding description) are encoded into vector representations using our fine-tuned
5, model; relevance is then scored using cosine similarity.
• Concurrently, for sparse retrieval, we utilize the BM25 algorithm to compute lexical similarity
scores between the query (title and description) and each corpus entry (title and description).</p>
        <p>The relevance scores from these two retrieval components are first normalized to a common range.
These normalized scores, denoted as 25, for BM25 and , for the vector-based semantic
similarity, are then integrated using a weighted linear fusion. The final fused score, , for each
candidate document is computed as:</p>
        <p>= (25 · 25,) + ( · ,)
Based on empirical evaluation on our validation set, we determined the optimal weights to be 25 =
0.15 for the BM25 component and  = 0.85 for the semantic vector component.</p>
        <p>
          The resulting candidate list, ranked by , is subsequently passed to the second stage, where
we used a fine-tuned version of jina-reranker-v2-base-multilingual cross-encoder model
released by Jina AI [
          <xref ref-type="bibr" rid="ref10">10</xref>
          ] to perform a more fine-grained relevance assessment on the top-10 candidates
to produce the final ranked output.
        </p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Experimental Setup</title>
      <sec id="sec-5-1">
        <title>5.1. Baseline</title>
        <p>
          We established a baseline using the paraphrase-multilingual-MiniLM-L12-v2 model [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ], which
was the oficial baseline for this task, providing a reference for measuring improvements.
        </p>
      </sec>
      <sec id="sec-5-2">
        <title>5.2. Embedding Model Selection (Zero-shot)</title>
        <p>
          We evaluated several pretrained multilingual embedding models in a
zeroshot setting on the validation data. The models tested included BGE-M3 [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ],
multilingual-E5-large, multilingual-E5-large-instruct [
          <xref ref-type="bibr" rid="ref9">9</xref>
          ], LaBSE [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ], and
paraphrase-multilingual-mpnet-base-v2 [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ]. The multilingual-E5-large-instruct
model demonstrated significantly superior performance, leading to its selection as our base model.
        </p>
      </sec>
      <sec id="sec-5-3">
        <title>5.3. Fine-tuning Approaches (Without Full Curriculum)</title>
        <p>The fine-tuning experiments utilized Multiple Negatives Ranking Loss (MNRL). The positive instances
for this loss consisted of pairs structured as either (job title, job title) or (job title, job title + description),
where the first element consistently served as the anchor and the second as the positive example. We
conducted preliminary fine-tuning experiments:
• Fine-tuning multilingual 5, on (job title, job title) pairs resulted in overfitting.
• Fine-tuning multilingual 5, on (job title, job title + description) pairs showed
substantial improvement over the title-only approach and zero-shot performance.</p>
        <p>
          We also experimented with ESCOXLM-R [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], a multilingual transformer model pretrained on ESCO data.
We fine-tuned the model on (job title, job title + description) pairs, testing mean, [CLS] token, and
attention pooling strategies for deriving sentence embeddings. All pooling strategies yielded similar
average mAP scores on the validation set.
        </p>
      </sec>
      <sec id="sec-5-4">
        <title>5.4. Curriculum Learning Configuration</title>
        <p>The cross-lingual curriculum learning strategy for the multilingual 5, model was
implemented in six stages. Throughout each curriculum stage, we also employed MNRL. The data composition
for each stage was as follows:
• Stage 1: 100% monolingual (job title, job title) pairs (en-en, es-es, de-de).
• Stage 2: 60% monolingual (job title, job title) pairs; 40% cross-lingual (job title, job title) pairs
(en-es, en-de).
• Stage 3: 60% monolingual (job title, job title); 20% cross-lingual (job title, job title); 20%
monolingual (job title, job title + description).
• Stage 4: 30% monolingual (job title, job title); 20% cross-lingual (job title, job title); 30%
monolingual (job title, job title + description); 20% cross-lingual (job title, job title + description).
• Stage 5: 10% monolingual (job title, job title); 10% cross-lingual (job title, job title); 60%
monolingual (job title, job title + description); 20% cross-lingual (job title, job title + description).
• Stage 6: 70% monolingual (job title, job title + description); 30% cross-lingual (job title, job title +
description).</p>
      </sec>
      <sec id="sec-5-5">
        <title>5.5. Reranker Fine-tuning</title>
        <p>For the second-stage refinement of our retrieval pipeline, we employed the
jina-reranker-v2-base-multilingual cross-encoder model. This model was specifically
ifne-tuned for the task using a Binary Cross-Entropy (BCE) loss function. The training data was
formulated from (query, candidate document, label) tuples. Each candidate document consisted of
a job title concatenated with its corresponding description. Positive instances were created using
known relevant query-document pairs ( = 1), while hard negative instances ( = 0) were
incorporated to improve the model’s discriminative power. These hard negatives were mined from the
corpus by utilizing the jina-embedding-v3 model [14] to retrieve documents (job titles and their
descriptions) that were highly ranked for a given query but were non-relevant.</p>
      </sec>
      <sec id="sec-5-6">
        <title>5.6. LLM-based Reranking</title>
        <p>We also explored using a Large Language Model for reranking. The top 10 documents retrieved by our
dense retrieval (curriculum-trained 5,) were presented to the Gemma3-27B model [15].
The model was prompted to reorder these candidates based on relevance to the query. The specific
prompt template utilized for this task is detailed in Listing 1:</p>
        <p>Listing 1: LLM Reranking Prompt Template
Below is a query and a list of 10 candidate job titles.</p>
        <p>Rank them from most relevant (rank=1) to least relevant (rank=10).</p>
        <p>Query: {query}
Candidates:
{candidates}
Return only the ordering as comma-separated numbers (e.g. 3,1,2,...):
In the prompt above, {query} is replaced with the actual query job title, and {candidates} is replaced
with a numbered list of the top-10 candidate job titles and their descriptions retrieved by the dense
model.</p>
      </sec>
    </sec>
    <sec id="sec-6">
      <title>6. Results and Discussion</title>
      <p>All results reported in this section are Mean Average Precision (mAP) scores obtained on the oficial
validation set, unless otherwise specified for test set evaluations.</p>
      <sec id="sec-6-1">
        <title>6.1. Zero-shot Embedding Model Performance</title>
        <p>To establish a baseline and select a strong foundation model, we evaluated several pretrained multilingual
embedding models in a zero-shot setting. Table 1 presents these results. The 5, model
achieved the highest average mAP of 47.27%, outperforming the oficial task baseline by approximately
9%. This superior performance justified its selection as our base model for fine-tuning.</p>
      </sec>
      <sec id="sec-6-2">
        <title>6.2. Performance of Fine-tuned Models (Without Full Curriculum)</title>
        <p>We then investigated the impact of fine-tuning on performance, initially without employing the full
curriculum learning strategy. Table 2 shows the results for 5, and ESCOXLM-R (with
various pooling strategies) when fine-tuned on pairs of (job title, job title + job description).
Finetuning 5, improved its average mAP from 47.27% (zero-shot) to 53.82%. This underscores
the value of incorporating job descriptions, which provide essential context for disambiguating job
titles. The 5, model also consistently outperformed the ESCOXLM-R variants, despite the
latter being pretrained specifically on ESCO data. We hypothesize that the base architecture and the
instruction-tuning of the E5 model make it more amenable for fine-tuning on this specific task structure.</p>
      </sec>
      <sec id="sec-6-3">
        <title>6.3. Impact of Curriculum Learning and Reranking Strategies</title>
        <p>Table 3 illustrates the performance of our system, which incorporates the full cross-lingual curriculum
learning strategy for 5,, and subsequently evaluates the impact of diferent reranking
5, (LLM reranking)
5, (BM25 &amp; cross-encoder reranking)
Language
mAP (dev)</p>
        <p>mAP (test)
English
German
Spanish
Average
English
German
Spanish
Average
English
German
Spanish
Average
approaches. The application of our cross-lingual curriculum learning strategy boosted the average
mAP of 5, on the development set to 55.76%, a substantial improvement from the 53.82%
achieved with direct fine-tuning (Table 2). Further enhancements were observed with reranking.
• The hybrid retrieval approach yielded the highest average mAP on the validation set (55.97%
mAP).
• LLM-based reranking with Gemma3-27B also showed a slight improvement over the
curriculumonly model on the validation set (55.82% mAP).</p>
        <p>On the test set, both the curriculum-trained model without further reranking and our system (curriculum
+ BM25 + cross-encoder reranker) achieved an average mAP of 53%. This indicates good generalization
for the curriculum-trained model, though the slight gains from reranking observed on the validation
set did not fully translate to the test set for all configurations.</p>
      </sec>
      <sec id="sec-6-4">
        <title>6.4. Impact of Job Descriptions at Inference</title>
        <p>A core component of our methodology is the enrichment of job titles with full descriptions, either
sourced from ESCO or generated by an LLM. To explicitly validate the necessity of this step during
inference, we conducted an ablation study on the validation set. We compared the performance of our
ifnal curriculum-trained model under two conditions: 1) the standard approach, where query titles are
used to retrieve from a corpus of job title + description, and 2) a title-only approach, where query titles
retrieve from a corpus containing only job titles, with descriptions omitted.</p>
        <p>The results, presented in Table 4, confirm the significant benefit of including descriptions. The average
mAP dropped from 55.76% to 53.18% when descriptions were removed from the inference corpus. This
performance decrease underscores that the contextual information provided by the descriptions is
crucial for the model to disambiguate similar or ambiguous job titles, justifying our data augmentation
strategy for the inference phase.
Curriculum E5 (Title + Description) 0.5226
Curriculum E5 (Title Only) 0.4992</p>
        <p>English</p>
      </sec>
      <sec id="sec-6-5">
        <title>6.5. Computational Considerations</title>
        <p>All experiments were conducted on a single NVIDIA A100 GPU. We analyzed the trade-of between
retrieval quality and computational cost for our primary system configurations. The baseline
curriculumtrained E5-large-instruct model ofers the lowest latency, requiring only a single embedding pass for the
query followed by a fast vector search. Our final system, which adds BM25 (negligible overhead) and a
cross-encoder reranker, introduces higher latency by requiring 10 additional forward passes per query.
This cost yielded a marginal mAP improvement (55.76% to 55.97% on validation), making the system
suitable for applications where accuracy is paramount. The LLM-based reranking using Gemma3-27B
was the most resource-intensive approach, incurring substantial latency and memory usage for a
minimal performance gain on the validation set and a drop on the test set, proving impractical for this
task.</p>
      </sec>
      <sec id="sec-6-6">
        <title>6.6. Limitations</title>
        <p>Despite the promising results, this study has certain limitations that should be acknowledged. Firstly,
our data augmentation strategy relies on LLM-generated descriptions for inference, and while eforts
were made to ensure contextual plausibility, the synthetic data may not fully capture the stylistic
diversity or factual nuances of authentic ESCO or real-world job descriptions. Discrepancies in quality
or representativeness could subtly influence model performance on job titles reliant on these synthetic
contexts. Secondly, the system’s evaluation is based on the English, Spanish, and German languages
within the TalentCLEF dataset. Its generalization to entirely diferent languages or highly specialized
job domains not well-represented in the training data requires further investigation; the observed lower
performance for German titles across several models may hint at existing cross-lingual representation
challenges. Thirdly, the current work did not incorporate specific mechanisms to assess or mitigate
potential performance disparities across job titles that may be stereotypically associated with
diferent gender groups; addressing this is an important consideration for ensuring fairness in practical
applications. Finally, the optimal configuration of various components, such as curriculum learning
parameters, hybrid retrieval weights, or the selected LLM prompts, was determined based on validation
set performance. As is common in such iterative development, this process may have led to some degree
of adaptation to the validation data’s specific characteristics, as suggested by the slight variations in
performance gains between the validation and test sets for some of our reranking configurations.</p>
      </sec>
    </sec>
    <sec id="sec-7">
      <title>7. Conclusion and Future Work</title>
      <p>Our participation in TalentCLEF 2025 Task A demonstrated the eficacy of a carefully designed
curriculum learning strategy combined with data augmentation, hybrid retrieval, and cross-encoder reranking
for multilingual job title matching. The phased introduction of complexity and cross-lingual signals
enabled our model to achieve strong performance. For future work, we plan to explore more sophisticated
hard negative mining techniques for both the embedding model and the reranker.</p>
      <p>Further investigation and experimenting with larger, more capable LLMs for reranking (perhaps
with more elaborate prompting) could also yield improvements. Additionally, we intend to improve
the synthetic data generation process by ensuring that the descriptions generated by the LLM more
accurately reflect the linguistic style, content scope, and structural properties inherent in authentic
ESCO descriptions.</p>
    </sec>
    <sec id="sec-8">
      <title>Declaration on Generative AI</title>
      <p>During the preparation of this work, the authors used Gemini 2.5 Pro in order to: Grammar and
spelling check. After using this tool, the authors reviewed and edited the content as needed and take
full responsibility for the publication’s content.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>L.</given-names>
            <surname>Gasco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Fabregat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>García-Sardiña</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Estrella</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Deniz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rodrigo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zbib</surname>
          </string-name>
          ,
          <article-title>Overview of the TalentCLEF 2025 Shared Task: Skill and Job Title Intelligence for Human Capital Management, in: International Conference of the Cross-Language Evaluation Forum for European Languages</article-title>
          , Springer,
          <year>2025</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>R.</given-names>
            <surname>Zbib</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. A.</given-names>
            <surname>Lacasa</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Retyk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Poves</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Aizpuru</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Fabregat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Simkus</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>García-Casademont</surname>
          </string-name>
          ,
          <article-title>Learning job titles similarity from noisy skill labels</article-title>
          ,
          <year>2023</year>
          . URL: https://arxiv.org/abs/2207.00494. arXiv:
          <volume>2207</volume>
          .
          <fpage>00494</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>N.</given-names>
            <surname>Laosaengpha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Tativannarat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Piansaddhayanon</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Rutherford</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Chuangsuwanich</surname>
          </string-name>
          ,
          <article-title>Learning job title representation from job description aggregation network</article-title>
          , in: L.
          <string-name>
            <surname>-W. Ku</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Martins</surname>
          </string-name>
          , V. Srikumar (Eds.),
          <source>Findings of the Association for Computational Linguistics ACL</source>
          <year>2024</year>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Bangkok, Thailand and virtual meeting,
          <year>2024</year>
          , pp.
          <fpage>1319</fpage>
          -
          <lpage>1329</lpage>
          . URL: https://aclanthology.org/
          <year>2024</year>
          .findings-acl.
          <volume>77</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , R. van der Goot, B. Plank,
          <string-name>
            <surname>ESCOXLM-R: Multilingual</surname>
          </string-name>
          taxonomy
          <article-title>-driven pre-training for the job market domain</article-title>
          ,
          <source>in: Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume</source>
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          ,
          <source>Association for Computational Linguistics</source>
          , Toronto, Canada,
          <year>2023</year>
          , pp.
          <fpage>11871</fpage>
          -
          <lpage>11890</lpage>
          . URL: https://aclanthology.org/
          <year>2023</year>
          .
          <article-title>acl-long</article-title>
          .
          <volume>662</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>D.</given-names>
            <surname>Deniz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Retyk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>García-Sardiña</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Fabregat</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Gasco</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zbib</surname>
          </string-name>
          ,
          <article-title>Combined unsupervised and contrastive learning for multilingual job recommendation</article-title>
          , in: M.
          <string-name>
            <surname>Kaya</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Bogers</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Graus</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Johnson</surname>
          </string-name>
          , J.
          <string-name>
            <surname>-J. Decorte</surname>
          </string-name>
          , T. D. Bie (Eds.),
          <source>Proceedings of the 4th Workshop on Recommender Systems for Human Resources (RecSys in HR 2024)</source>
          , volume
          <volume>3788</volume>
          <source>of CEUR Workshop Proceedings</source>
          , CEUR-WS.org, Bari, Italy,
          <year>2024</year>
          , pp.
          <fpage>1</fpage>
          -
          <lpage>8</lpage>
          . URL: https://ceur-ws.
          <source>org/</source>
          Vol-
          <volume>3788</volume>
          /RecSysHR2024-paper_ 3.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>L.</given-names>
            <surname>Gascó</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F. M.</given-names>
            <surname>Hermenegildo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.-S.</given-names>
            <surname>Laura</surname>
          </string-name>
          , D. C. Daniel, P. Estrella,
          <string-name>
            <given-names>R.</given-names>
            <surname>Alvaro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Rabih</surname>
          </string-name>
          ,
          <article-title>Talentclef 2025 corpus: Skill and job title intelligence for human capital management, 2025</article-title>
          . URL: https: //doi.org/10.5281/zenodo.15240844. doi:
          <volume>10</volume>
          .5281/zenodo.15240844.
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>M.</given-names>
            <surname>le Vrang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Papantoniou</surname>
          </string-name>
          , E. Pauwels,
          <string-name>
            <given-names>P.</given-names>
            <surname>Fannes</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Vandensteen</surname>
          </string-name>
          , J. De Smedt,
          <article-title>Esco: Boosting job matching in europe with semantic interoperability</article-title>
          ,
          <source>Computer</source>
          <volume>47</volume>
          (
          <year>2014</year>
          )
          <fpage>57</fpage>
          -
          <lpage>64</lpage>
          . doi:
          <volume>10</volume>
          .1109/
          <string-name>
            <surname>MC</surname>
          </string-name>
          .
          <year>2014</year>
          .
          <volume>283</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>A.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Hui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Gao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Lv</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Hu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Ge</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Tu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Dang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Bao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Xue</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Q.</given-names>
            <surname>Zhu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Men</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Gao</surname>
          </string-name>
          , S. Liu,
          <string-name>
            <given-names>S.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Tang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Yin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Ren</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Fan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Su</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Wan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Cui</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhou</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Qiu</surname>
          </string-name>
          ,
          <source>Qwen3 technical report</source>
          ,
          <year>2025</year>
          . URL: https://arxiv.org/abs/2505.09388. arXiv:
          <volume>2505</volume>
          .
          <fpage>09388</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Huang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Majumder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Wei</surname>
          </string-name>
          ,
          <source>Multilingual e5 text embeddings: A technical report</source>
          ,
          <year>2024</year>
          . URL: https://arxiv.org/abs/2402.05672. arXiv:
          <volume>2402</volume>
          .
          <fpage>05672</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>M.</given-names>
            <surname>Günther</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Ong</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Mohr</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Abdessalem</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Abel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. K.</given-names>
            <surname>Akram</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Guzman</surname>
          </string-name>
          , G. Mastrapas,
          <string-name>
            <given-names>S.</given-names>
            <surname>Sturua</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Werk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Xiao</surname>
          </string-name>
          , Jina embeddings
          <volume>2</volume>
          :
          <fpage>8192</fpage>
          -token
          <article-title>general-purpose text embeddings for long documents</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2310</volume>
          .
          <fpage>19923</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>N.</given-names>
            <surname>Reimers</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Gurevych</surname>
          </string-name>
          ,
          <article-title>Sentence-bert: Sentence embeddings using siamese bert-networks</article-title>
          ,
          <source>in: Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics</source>
          ,
          <year>2019</year>
          . URL: http://arxiv.org/abs/
          <year>1908</year>
          .10084.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>J.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Xiao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Zhang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lian</surname>
          </string-name>
          ,
          <string-name>
            <surname>Z</surname>
          </string-name>
          . Liu, M3
          <article-title>-embedding: Multi-linguality, multifunctionality, multi-granularity text embeddings through self-knowledge distillation</article-title>
          , in: L.
          <string-name>
            <surname>-W. Ku</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Martins</surname>
          </string-name>
          , V. Srikumar (Eds.),
          <source>Findings of the Association for Computational Linguistics: ACL</source>
          <year>2024</year>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Bangkok, Thailand,
          <year>2024</year>
          , pp.
          <fpage>2318</fpage>
          -
          <lpage>2335</lpage>
          . URL: https://aclanthology.org/
          <year>2024</year>
          .findings-acl.
          <volume>137</volume>
          /. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2024</year>
          .findings-acl.
          <volume>137</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>F.</given-names>
            <surname>Feng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Cer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Arivazhagan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>Language-agnostic BERT sentence embedding</article-title>
          , in: S. Muresan,
          <string-name>
            <given-names>P.</given-names>
            <surname>Nakov</surname>
          </string-name>
          ,
          <string-name>
            <surname>A</surname>
          </string-name>
          . Villavicencio (Eds.),
          <source>Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume</source>
          <volume>1</volume>
          :
          <string-name>
            <surname>Long</surname>
            <given-names>Papers)</given-names>
          </string-name>
          , Association for Computational
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>