<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Identifying Semantic Relationships Between Research Topics Using Large Language Models in a Zero-Shot Learning Setting</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Tanay Aggarwal</string-name>
          <email>tanay.aggarwal@open.ac.uk</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Angelo Salatino</string-name>
          <email>angelo.salatino@open.ac.uk</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Francesco Osborne</string-name>
          <email>francesco.osborne@open.ac.uk</email>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Enrico Motta</string-name>
          <email>enrico.motta@open.ac.uk</email>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Business and Law, University of Milano Bicocca</institution>
          ,
          <addr-line>Milan, IT</addr-line>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Knowledge Media Institute, The Open University</institution>
          ,
          <addr-line>Milton Keynes</addr-line>
          ,
          <country country="UK">UK</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Zero-Shot Learning, Large Language Models, Ontology Generation, Research Topics</institution>
          ,
          <addr-line>Scholarly Knowl-</addr-line>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2024</year>
      </pub-date>
      <fpage>0000</fpage>
      <lpage>0003</lpage>
      <abstract>
        <p>Knowledge Organization Systems (KOS), such as ontologies, taxonomies, and thesauri, play a crucial role in organising scientific knowledge. They help scientists navigate the vast landscape of research literature and are essential for building intelligent systems such as smart search engines, recommendation systems, conversational agents, and advanced analytics tools. However, the manual creation of these KOSs is costly, time-consuming, and often leads to outdated and overly broad representations. As a result, researchers have been exploring automated or semi-automated methods for generating ontologies of research topics. This paper analyses the use of large language models (LLMs) to identify semantic relationships between research topics. We specifically focus on six open and lightweight LLMs (up to 10.7 billion parameters) and use two zero-shot reasoning strategies to identify four types of relationships: broader, narrower, same-as, and other. Our preliminary analysis indicates that Dolphin2.1-OpenOrca-7B performs strongly in this task, achieving a 0.853 F1-score against a gold standard of 1,000 relationships derived from the IEEE Thesaurus. These promising results bring us one step closer to the next generation of tools for automatically curating KOSs, ultimately making the scientific literature easier to explore.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>CEUR
ceur-ws.org</p>
    </sec>
    <sec id="sec-2">
      <title>1. Introduction</title>
      <p>
        Knowledge Organization Systems (KOS), like ontologies, taxonomies, and thesauri, designed for
research topics, are instrumental for structuring, managing, and retrieving information from
digital libraries, enabling eficient knowledge discovery [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. Major publishers like ACM, IEEE,
PubMed, and SpringerNature employ KOS like the ACM Computing Classification System 1
their published content [
        <xref ref-type="bibr" rid="ref2 ref3 ref4">2, 3, 4</xref>
        ]. KOSs play also a crucial role in enabling intelligent systems
to navigate and interpret academic literature efectively [
        <xref ref-type="bibr" rid="ref5 ref6">5, 6</xref>
        ], serving as the foundation for
tools like search engines [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ], conversational agents [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], analytics dashboards [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], and academic
recommender systems [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]. Additionally, KOS provide a robust representation of research topics,
crucial for various AI-driven analyses of scientific literature [
        <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
        ].
      </p>
      <p>
        However, maintaining these KOSs is a growing challenge. The rapid expansion of scientific
literature, with an estimated 2.5 million new papers published annually [
        <xref ref-type="bibr" rid="ref11">11</xref>
        ], necessitate
continuous curation to reflect the latest advancements. Besides, manually curating them has become
increasingly costly and time-consuming, highlighting the need for innovative solutions to keep
pace with the evolving scientific landscape.
      </p>
      <p>
        In contrast, the emergence of Large Language Models (LLMs) has revolutionised the field of
Artificial Intelligence (AI), enabling deeper language comprehension, including grasping the
semantic of word and sentences, inferring relationships between concepts, resolving ambiguities,
and understanding overall text meaning [
        <xref ref-type="bibr" rid="ref12">12</xref>
        ].
      </p>
      <p>Given the challenges in maintaining up-to-date ontologies of research areas, and the recent
advancements in AI, this paper presents an analysis showcasing the capabilities of LLMs in
identifying semantic relationships between pairs of research topics. Specifically, we focus on
six open and lightweight models (up to 10.7 billion parameters), and we employ two zero-shot
reasoning strategies to identify four relationships types (e.g., broader, narrower, same-as, and
other ), on a gold standard of 1,000 relationship (250 per relationship).</p>
      <p>Our main objective is to develop an innovative pipeline to automatically generate and maintain
ontologies of research topics. This pipeline will allow us to curate existing ontologies, create
more granular representations of research concepts, and expand the development of ontologies
into other scientific fields. In this context, our research focuses on determining whether LLMs
can efectively aid in this process.</p>
      <p>In brief, this paper contributes to the literature with a preliminary analysis of whether
zeroshot reasoning can efectively identify semantic relationships between research topics using
open, smaller models, with a focus on sustainability. The gold standard and the code we used to
run our experiments are available on a GitHub repository2.</p>
      <p>The remainder of this paper is structured as follows. Section 2 provides an overview of the
literature. Section 3 describes the dataset and the approach we have devised to conduct our
experiments. Section 4 presents and discusses our results. Finally, in Section 5 we conclude the
paper and provide future directions.</p>
    </sec>
    <sec id="sec-3">
      <title>2. Literature Review</title>
      <p>In this section, we review the literature focusing on two key aspects relevant to our work:
ontologies of research areas and automatic ontology generation.</p>
      <sec id="sec-3-1">
        <title>2.1. Ontologies of research areas</title>
        <p>
          In the literature there are several ontologies, or more in general KOSs, within the scientific
ecosystem, which can support the exploration process in digital libraries, the production of
scholarly analytics, and modelling research dynamics [
          <xref ref-type="bibr" rid="ref13 ref2">2, 13</xref>
          ]. These include Medical Subject
2Gold Standard and Code for experiments - https://github.com/ImTanay/LLM-Semantic-Relationship-Analysis
Headings (MeSH) [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ], ACM Computing Classification System (CCS), the Computer Science
Ontology (CSO) [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ], AGROVOC, Mathematical Subject Classification (MSC) [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ], and Physics
Subject Headings (PhySH) [
          <xref ref-type="bibr" rid="ref14">14</xref>
          ]. MeSH is a comprehensive controlled vocabulary, with more
than 30K concepts, developed and maintained by the National Library of Medicine [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ]. It is
widely used in the medical and health sciences, and it receives yearly updates. The ACM
Computing Classification System 3 is a taxonomy of research topics in the field of Computer
Science, covering about 2K research topics. It is curated by ACM, the world’s largest educational
and scientific computing society, and the last update dates back to 2012. CSO is the largest
ontology of research topic in Computer Science, covering 14K research areas [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ]. It has been
automatically generated using the Klink-2 algorithm [
          <xref ref-type="bibr" rid="ref15">15</xref>
          ] on a dataset of 16 million scientific
articles, and receives yearly updates. The IEEE Thesaurus mainly covers the field of Engineering
but also contains diferent concepts relevant to Computer Science. It is curated by the Institute of
Electrical and Electronics Engineers4. It contains around 5.6K topics and 24K relationships, and
receives yearly updates. MSC is a comprehensive taxonomy with over 6.5K concepts, covering
a wide range of mathematical disciplines, from pure mathematics to applied mathematics and
statistics. It is curated by the American Mathematical Society and zbMATH, and receives
updates every 10 years [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ].
        </p>
        <p>
          Most of the existing KOSs are curated manually, usually by a committee of domain experts
who periodically meet to discuss the updates for the next version. This process, however, makes
ontologies evolve slowly and hence prone to becoming outdated. Additionally, such a manual
curation is costly and increasingly unsustainable given the rapid rate at which new research is
published [
          <xref ref-type="bibr" rid="ref11 ref13">11, 13</xref>
          ]. Automating the creation of research area ontologies can overcome existing
limitations by ensuring these ontologies remain current. This, in turn, can enhance cataloguing,
retrievability, and the various downstream applications mentioned earlier.
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>2.2. Automatic Ontology Generation</title>
        <p>
          The automatic generation of ontologies is a field of research whose objective is to overcome
the challenges of traditional ontology creation, hence making it more scalable and eficient.
Techniques that are usually employed include natural language processing, clustering techniques,
or statistical methods [
          <xref ref-type="bibr" rid="ref15 ref16 ref17">16, 17, 15</xref>
          ]. For instance, Text2Onto [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ] is a tool that can create ontologies
from a corpus of documents. This method identifies synonyms, sub-/superclass hierarchies, and
through the application of NLP techniques it can learn hierarchical structures between terms,
leveraging phrases such as “such as...” and “and other...”.
        </p>
        <p>
          Shan et al. [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ] used a similar method to develop Fields of Study for Microsoft Academic,
combining manually created concepts with topics derived from Wikidata. However, this
approach relied heavily on Wikidata and did not utilize metadata associated with research papers.
The OpenAlex team [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] also employed a similar strategy, expanding upon the “All Science
Journal Classification” structure in Scopus and incorporating topics extracted from papers using
citation analysis.
        </p>
        <p>
          Some studies have explored a hybrid approach, combining ontology learning with
crowdsourcing to integrate statistical measures and user feedback [
          <xref ref-type="bibr" rid="ref20 ref21">20, 21</xref>
          ]. Specifically, human support
        </p>
        <sec id="sec-3-2-1">
          <title>3The ACM Computing Classification System – http://www.acm.org/publications/class-2012</title>
          <p>
            4IEEE Taxonomy - https://www.ieee.org/content/dam/ieee-org/ieee/web/org/pubs/ieee-taxonomy.pdf
has been integrated to evaluate an automatically generated ontology [
            <xref ref-type="bibr" rid="ref20">20</xref>
            ].
          </p>
          <p>
            More recently, researchers have begun to leverage LLMs for the creation of taxonomies,
ontologies, and knowledge graphs [
            <xref ref-type="bibr" rid="ref22">22</xref>
            ]. For instance, Chen et al. [
            <xref ref-type="bibr" rid="ref23">23</xref>
            ] proposed a two-module
approach for taxonomy generation. The first module predicts parent-child relationships, and
the second module reconciles these predictions into tree structures. The model is trained on
subtrees from Wordnet and evaluated on separate Wordnet subtrees.
          </p>
          <p>Given the new opportunities unlocked by LLMs, we hypothesise that significant progress
can be made to tackle the ongoing challenges of curating ontologies of research areas.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>3. Material and Methods</title>
      <sec id="sec-4-1">
        <title>This section describes the task, the gold standard, and the employed LLMs.</title>
        <sec id="sec-4-1-1">
          <title>3.1. Task definition and experiments</title>
          <p>The task is to classify the semantic relationship holding between pairs of research topics (  ,   )
according to four categories which are essential for ontology generation. More formally, this is
single-label multi-class classification problem, and the categories are:
• broader :   is a parent topic of   . E.g., ontological languages is a broader area than owl
• narrower :   is a child topic of   . E.g., nosql is a specific area within databases
• same-as:   and   can be used interchangeably to refer to same concept. E.g., haptic
interface and haptic device
• other :   and   do not relate according to the above categories. E.g., blockchain and user
interfaces
Our experiments consist of two zero-shot reasoning strategies, as depicted in Fig. 1.</p>
          <p>One-way Strategy: This experiment, highlighted by the red dashed box in Fig. 1, involves
taking each pair of research topics, generating a prompt using a specially designed template
(see Appendix A), submitting it to the LLM, and then analysing the response to determine the
appropriate classification.</p>
          <p>The prompt template is identical for both strategies and all models. It was carefully refined
through an iterative process to ensure optimal performance and consistency across all models.</p>
          <p>Two-way Strategy: This experiment, highlighted by the green dashed box in Fig. 1, involves
running the one-way strategy twice. First, we identify the relationship between topics   and
  . Then, in a fresh context, it identifies the relationship when the topics are swapped. This is
possible because the relationships broader and narrower are inverses of each other, same-as is
symmetric, and by definition also other is symmetric.</p>
          <p>Finally, we set empirical rules (cyan box in Fig. 1) to mitigate the agreement/disagreement
between the two branches of the two-way strategy. These rules are designed with the aim of
prioritising the development of the hierarchical structure. Let f(X) and s(X) represent the
relationship types returned by the first and second branches of our two-way strategy, respectively.
Additionally, let len(  ) denote the length of the topic’s surface form. We defined the rules as
follow:
1. broader :- f(broader) ∧ s(narrower)
2. narrower :- f(narrower) ∧ s(broader)
3. broader :- ((f(narrower) ∧ s(narrower)) ∨ (f(broader) ∧ s(broader))) ∧ len(  ) ≤ len(  )
4. narrower :- ((f(narrower) ∧ s(narrower)) ∨ (f(broader) ∧ s(broader))) ∧ len(  ) &gt; len(  )
5. same-as :- f(same-as) ∧ s(same-as)
6. broader :- (f(broader) ∧ s(other)) ∨ (f(other) ∧ s(narrower))
7. narrower :- (f(narrower) ∧ s(other)) ∨ (f(other) ∧ s(broader))
8. :- f(X)</p>
        </sec>
        <sec id="sec-4-1-2">
          <title>3.2. Gold standard</title>
          <p>
            As gold standard, we selected a balanced sample of 1,000 semantic relationships from the IEEE
Thesaurus, with 250 relationships per category. The IEEE Thesaurus includes 11,570 descriptive
terms in the field of Engineering, and serves as a standardised vocabulary of technical terms
for indexing and retrieving content within the IEEE digital library. Specifically, we utilised the
IEEE Thesaurus v1.025, which is dated to July 2023, and is available as a PDF file following the
ANSI/NISO Z39.4-2021 standard [
            <xref ref-type="bibr" rid="ref24">24</xref>
            ].
          </p>
          <p>To generate our gold standard, we first extracted the hierarchical structure and relationships
between terms from the original IEEE Thesaurus PDF document and transformed it into RDF
5A copy of version 1.02 of the IEEE Thesaurus is available at https://github.com/angelosalatino/
ieee-taxonomy-thesaurus-rdf/blob/main/source/ieee-thesaurus_2023.pdf.
format6. We represented the various relationships according to the Simple Knowledge
Organization System (SKOS) notation: i) skos:broader, ii) skos:narrower, iii) skos:altLabel,
iv) skos:prefLabel, v) skos:related.</p>
          <p>We randomly sampled 250 relationships each for the categories broader, narrower, and same-as.
For the latter, we used a combination of skos:altLabel and skos:prefLabel. Finally, we
randomly coupled topics to generate 250 other relationships, ensuring that these new pairs did
not share any of the previously established semantic relationships within the IEEE Thesaurus.</p>
        </sec>
        <sec id="sec-4-1-3">
          <title>3.3. Large Language Models</title>
          <p>For our experiments we selected six quantised LLMs, that had been quantised to 8-bit precision.
These included four fine-tuned versions of Mistral-7B, in addition to SOLAR and LLaMa 3.</p>
          <p>
            Dolphin-2.1-Mistral-7B: (shortened as dolphin-mistral7) is a decoder-only model with 7
billion parameters and a token context capacity of 4096. It is based on Mistral-7B and fine-tuned
with the Dolphin8 dataset, which is an open-source implementation of Microsoft’s Orca [
            <xref ref-type="bibr" rid="ref25">25</xref>
            ],
with an addition of Airoboros9 dataset.
          </p>
          <p>
            Dolphin-2.6-Mistral-7B-dpo-laser: (shortened as dolphin-mistral-dpo10) is based on
Mistal7B and fine-tuned on top of Dolphin DPO using Layer Selective Rank Reduction (LASER) [
            <xref ref-type="bibr" rid="ref26">26</xref>
            ].
It is a decoder-only model with 7 billion parameters, ofering a context window of 4096 tokens.
          </p>
          <p>
            Dolphin2.1-OpenOrca-7B: (shortened as dolphin-openorca11) is a model that blends
Dolphin2.1-Mistral-7B and Mistral-7B-OpenOrca12 models. These models were merged using the “ties
merge” [
            <xref ref-type="bibr" rid="ref27">27</xref>
            ] technique, keeping the same number of training parameters and token context
window size, respectively 7 billion and 4096.
          </p>
          <p>
            OpenChat-3.5-0106-Gemma: (shortened as openchat-gemma13) is a model trained on
openchat-3.5-0106 14 data using Conditioned Reinforcement Learning Fine-Tuning (C-RLFT)
framework [
            <xref ref-type="bibr" rid="ref28 ref29">28, 29</xref>
            ]. This model shares the same properties as openchat-3.5-0106, which is a
decoder-only model fine-tuned 15 on top of Mistral-7B. It consists of 7 billion parameters with a
context window size of 8192 tokens.
          </p>
          <p>
            SOLAR-10.7B-Instruct-v1.0: (shortened as solar 16) is a 10.7 billion parameters model with a
Depth Up-Scaling (DUS) architecture, which includes architectural modifications and continued
pretraining [
            <xref ref-type="bibr" rid="ref30">30</xref>
            ].
          </p>
          <p>
            Llama-3-8B-Instruct : (shortened as llama-317) is an auto-regressive model based on
transformer architecture. The updated versions of this model utilises supervised fine-tuning (SFT)
6The code we employed for converting the IEEE Thesaurus in RDF is available here https://github.com/angelosalatino/
ieee-taxonomy-thesaurus-rdf.
7Dolphin-2.1-Mistral-7B - https://huggingface.co/TheBloke/dolphin-2.1-mistral-7B-GGUF
8Dolphin Dataset - https://huggingface.co/datasets/cognitivecomputations/dolphin
9Airoboros - https://huggingface.co/datasets/jondurbin/airoboros-2.1
10Dolphin-2.6-Mistral-7B-dpo-laser - https://huggingface.co/TheBloke/dolphin-2.6-mistral-7B-dpo-laser-GGUF
11Dolphin2.1-OpenOrca-7B - https://huggingface.co/TheBloke/Dolphin2.1-OpenOrca-7B-GGUF
12Mistral-7B-OpenOrca - https://huggingface.co/Open-Orca/Mistral-7B-OpenOrca
13OpenChat-3.5-0106-Gemma - https://huggingface.co/gguf/openchat-3.5-0106-gemma-GGUF
14openchat-3.5-0106 - https://huggingface.co/openchat/openchat-3.5-0106
15Huggingface - https://huggingface.co/openchat/openchat-3.5-1210/discussions/4#658288f1168803bdee13d6b3
16SOLAR-10.7B-Instruct-v1.0 - https://huggingface.co/TheBloke/SOLAR-10.7B-Instruct-v1.0-GGUF
17Llama-3-8B-Instruct - https://huggingface.co/lmstudio-community/Meta-Llama-3-8B-Instruct-GGUF
and reinforcement learning with human feedback (RLHF). This model has 8 billion parameters
and a context length of 8k tokens [
            <xref ref-type="bibr" rid="ref31">31</xref>
            ].
          </p>
          <p>All these models are openly available on Huggingface. To run them, we used Google
Colaboratory, equipped with Nvidia’s V100 and L4 GPU(s), and used KoboldCpp18, which is a software
tool to operate with LLMs. KoboldCpp is a standalone program built on llama.cpp that makes
models accessible via an API endpoint.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>4. Results and Discussion</title>
      <p>We assessed the performance of the LLMs using precision, recall, and F1-score. Table 1 reports
the results of the one-way strategy. All models excel in precision (&gt;= 0.85) for the “same-as”
relationships, although most struggle with recall, except for the solar model (recall = 0.784).
The models also demonstrate high precision for “narrower” and high recall for “broader”. For
the “other” relationship, llama-3 and openchat-gemma exhibit high precision, dolphin-mistral
and dolphin-mistral-dpo exhibit high recall, whereas dolphin-openorca excels in both precision
and recall.</p>
      <p>Notably, dolphin-openorca emerges as the top performer with an average precision of 0.780,
recall of 0.733, and F1-score of 0.724. Indeed, dolphin-openorca has high precision for “narrower”,
“same-as”, and “other”, and high recall for “broader” and “other”.
The two-way strategy yielded dramatic improvements in average precision, recall, and
F1score, as reported in Table 2. Specifically, the values of precision for “broader” and recall
for “narrower” are significantly higher, efectively addressing the weaknesses of the one-way
strategy. Once again, dolphin-openorca emerges as the top performer with an impressive
average F1-score of 0.853, due to consistently high scores across all relationship types: 0.841
(broader), 0.829 (narrower), 0.845 (same-as), and 0.897 (other).</p>
      <p>Table 3 demonstrates that by exploiting the symmetric nature of the analysed semantic
relationships, the two-way strategy leads to a considerable improvement in F1-scores for most
models. This improvement is approximately 7% for solar and openchat-gemma, whereas for
the dolphin family models it exceeds 10%, with dolphin-mistral-dpo reaching a notable 24.3%
18KoboldCpp - https://github.com/LostRuins/koboldcpp
increase. In contrast, llama-3 appears to be the only model not significantly impacted by the
choice of strategy.</p>
    </sec>
    <sec id="sec-6">
      <title>5. Conclusions and Future Work</title>
      <p>In this paper, we evaluated the ability of six LLMs in identifying the semantic relationships
between research concepts, comparing their performance to a gold standard of 1,000
relationships from the IEEE Thesaurus. Our results demonstrate that state-of-the-art models like
dolphin-openorca achieve excellent zero-shot reasoning performance (0.853 of F1-score).</p>
      <p>Our future work will focus on several directions. Currently, we are expanding our analysis
to include additional models, such as non-quantised models, models with larger numbers of
parameters (e.g., LLaMa 3 70B), and proprietary models like ChatGPT 4.0 and the Claude
family. Furthermore, we plan to investigate fine-tuning some models (e.g., Google’s Gemma),
incorporate a reasoner into our pipeline to identify and resolve inconsistencies, and extend our
analysis to other scientific fields like Material Science, Medicine, and Physics.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgments</title>
      <p>We would like to thank Alessia Pisu, PhD Student from the University of Cagliari (IT) who
helped us in converting the IEEE Thesaurus from PDF to RDF.
For consistency we applied the same prompt across all the models. We engineered this prompt
through various refinements, to ensure optimal comprehension of the task and accurate
responses.</p>
      <p>Below is the template of our prompt, customised for each topic pair by substituting [TOPIC-A]
for the first topic and [TOPIC-B] for the second.</p>
      <p>Classify the relationship between '[TOPIC-A]' and '[TOPIC-B]' by applying
↪ the following relationship definitions:
1. '[TOPIC-A]' is-broader-than '[TOPIC-B]' if '[TOPIC-A]' is a
↪ super-category of '[TOPIC-B]', that is '[TOPIC-B]' is a type, a branch,
↪ or a specialized aspect of '[TOPIC-A]' or that '[TOPIC-B]' is a tool or
↪ a methodology mostly used in the context of '[TOPIC-A]' (e.g., car
↪ is-broader-than wheel).
2. '[TOPIC-A]' is-narrower-than '[TOPIC-B]' if '[TOPIC-A]' is a sub-category
↪ of '[TOPIC-B]', that is '[TOPIC-A]' is a type, a branch, or a
↪ specialized aspect of '[TOPIC-B]' or that '[TOPIC-A]' is a tool or a
↪ methodology mostly used in the context of '[TOPIC-B]' (e.g., wheel
↪ is-narrower-than car).
3. '[TOPIC-A]' is-same-as-than '[TOPIC-B]' if '[TOPIC-A]' and '[TOPIC-B]'
↪ are synonymous terms denoting an identical concept (e.g., beautiful
↪ is-same-as-than attractive), including when one is the plural form of
↪ the other (e.g., cat is-same-as-than cats).
4. '[TOPIC-A]' is-other-than '[TOPIC-B]' if '[TOPIC-A]' and '[TOPIC-B]'
↪ either have no direct relationship or share a different kind of
↪ relationship that does not fit into the other defined relationships.
Given the previous definitions, determine which one of the following
↪ statements is correct:
1. '[TOPIC-A]' is-broader-than '[TOPIC-B]'
2. '[TOPIC-B]' is-narrower-than '[TOPIC-A]'
3. '[TOPIC-A]' is-narrower-than '[TOPIC-B]'
4. '[TOPIC-B]' is-broader-than '[TOPIC-A]'
5. '[TOPIC-A]' is-same-as-than '[TOPIC-B]'
6. '[TOPIC-A]' is-other-than '[TOPIC-B]'
Answer by only stating the correct statement and its number.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>A.</given-names>
            <surname>Salatino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Aggarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mannocci</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Osborne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Motta</surname>
          </string-name>
          ,
          <article-title>A survey on knowledge organization systems of research fields: Resources and challenges</article-title>
          ,
          <source>arXiv preprint arXiv:2409.04432</source>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>E.</given-names>
            <surname>Dunne</surname>
          </string-name>
          , K. Hulek,
          <source>Mathematics subject classification</source>
          <year>2020</year>
          ,
          <source>EMS Newsletter</source>
          <year>2020</year>
          -
          <volume>3</volume>
          (
          <year>2020</year>
          )
          <fpage>5</fpage>
          -
          <lpage>6</lpage>
          . URL: http://dx.doi.org/10.4171/NEWS/115/2. doi:
          <volume>10</volume>
          .4171/news/115/2.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>C. E.</given-names>
            <surname>Lipscomb</surname>
          </string-name>
          ,
          <article-title>Medical subject headings (mesh</article-title>
          ),
          <source>Bulletin of the Medical Library Association</source>
          <volume>88</volume>
          (
          <year>2000</year>
          )
          <fpage>265</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>B.</given-names>
            <surname>Rous</surname>
          </string-name>
          ,
          <article-title>Major update to acm's computing classification system</article-title>
          ,
          <source>Communications of the ACM</source>
          <volume>55</volume>
          (
          <year>2012</year>
          )
          <fpage>12</fpage>
          -
          <lpage>12</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>J.</given-names>
            <surname>Beel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Gipp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Langer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Breitinger</surname>
          </string-name>
          ,
          <article-title>Paper recommender systems: a literature survey</article-title>
          ,
          <source>International Journal on Digital Libraries</source>
          <volume>17</volume>
          (
          <year>2016</year>
          )
          <fpage>305</fpage>
          -
          <lpage>338</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>M.</given-names>
            <surname>Gusenbauer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N. R.</given-names>
            <surname>Haddaway</surname>
          </string-name>
          ,
          <article-title>Which academic search systems are suitable for systematic reviews or meta-analyses? evaluating retrieval qualities of google scholar, pubmed, and 26 other resources</article-title>
          ,
          <source>Research synthesis methods 11</source>
          (
          <year>2020</year>
          )
          <fpage>181</fpage>
          -
          <lpage>217</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>A.</given-names>
            <surname>Meloni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Angioni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Salatino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Osborne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. R.</given-names>
            <surname>Recupero</surname>
          </string-name>
          , E. Motta,
          <article-title>Integrating conversational agents and knowledge graphs within the scholarly domain</article-title>
          ,
          <source>Ieee Access</source>
          <volume>11</volume>
          (
          <year>2023</year>
          )
          <fpage>22468</fpage>
          -
          <lpage>22489</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>S.</given-names>
            <surname>Angioni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Salatino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Osborne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. R.</given-names>
            <surname>Recupero</surname>
          </string-name>
          ,
          <string-name>
            <surname>E. Motta,</surname>
          </string-name>
          <article-title>The aida dashboard: a web application for assessing and comparing scientific conferences</article-title>
          ,
          <source>IEEE Access 10</source>
          (
          <year>2022</year>
          )
          <fpage>39471</fpage>
          -
          <lpage>39486</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>J. W.</given-names>
            <surname>Goodell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. M.</given-names>
            <surname>Lim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Pattnaik</surname>
          </string-name>
          ,
          <article-title>Artificial intelligence and machine learning in finance: Identifying foundations, themes, and research clusters from bibliometric analysis</article-title>
          ,
          <source>Journal of Behavioral and Experimental Finance</source>
          <volume>32</volume>
          (
          <year>2021</year>
          )
          <fpage>100577</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>A.</given-names>
            <surname>Salatino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Angioni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Osborne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. R.</given-names>
            <surname>Recupero</surname>
          </string-name>
          , E. Motta,
          <article-title>Diversity of expertise is key to scientific impact: a large-scale analysis in the field of computer science</article-title>
          ,
          <source>arXiv preprint arXiv:2306.15344</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>L.</given-names>
            <surname>Bornmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Mutz</surname>
          </string-name>
          ,
          <article-title>Growth rates of modern science: A bibliometric analysis based on the number of publications and cited references</article-title>
          ,
          <source>Journal of the Association for Information Science and Technology</source>
          <volume>66</volume>
          (
          <year>2015</year>
          )
          <fpage>2215</fpage>
          -
          <lpage>2222</lpage>
          . doi:https://doi.org/10.1002/asi.23329.
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>T.</given-names>
            <surname>Kojima</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. S.</given-names>
            <surname>Gu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Reid</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Matsuo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Iwasawa</surname>
          </string-name>
          ,
          <article-title>Large language models are zero-shot reasoners</article-title>
          ,
          <year>2023</year>
          . URL: https://arxiv.org/abs/2205.11916. arXiv:
          <volume>2205</volume>
          .
          <fpage>11916</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>A. A.</given-names>
            <surname>Salatino</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Thanapalasingam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mannocci</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Osborne</surname>
          </string-name>
          ,
          <string-name>
            <surname>E. Motta,</surname>
          </string-name>
          <article-title>The computer science ontology: a large-scale taxonomy of research areas, in: The Semantic Web-ISWC</article-title>
          <year>2018</year>
          : 17th International Semantic Web Conference, Monterey, CA, USA, October 8-
          <issue>12</issue>
          ,
          <year>2018</year>
          , Proceedings,
          <source>Part II 17</source>
          , Springer,
          <year>2018</year>
          , pp.
          <fpage>187</fpage>
          -
          <lpage>205</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>A.</given-names>
            <surname>Smith,</surname>
          </string-name>
          <article-title>Physics subject headings (physh)</article-title>
          ,
          <source>KO KNOWLEDGE ORGANIZATION 47</source>
          (
          <year>2020</year>
          )
          <fpage>257</fpage>
          -
          <lpage>266</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>F.</given-names>
            <surname>Osborne</surname>
          </string-name>
          , E. Motta,
          <article-title>Klink-2: Integrating multiple web sources to generate semantic topic networks</article-title>
          , in: M.
          <string-name>
            <surname>Arenas</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <string-name>
            <surname>Corcho</surname>
            , E. Simperl,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Strohmaier</surname>
            , M. d'Aquin,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Srinivas</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Groth</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Dumontier</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Heflin</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Thirunarayan</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Thirunarayan</surname>
          </string-name>
          , S. Staab (Eds.),
          <source>The Semantic Web - ISWC 2015</source>
          , Springer International Publishing, Cham,
          <year>2015</year>
          , pp.
          <fpage>408</fpage>
          -
          <lpage>424</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>P.</given-names>
            <surname>Cimiano</surname>
          </string-name>
          , J. Völker, Text2onto, in: A.
          <string-name>
            <surname>Montoyo</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Muńoz</surname>
          </string-name>
          , E. Métais (Eds.),
          <source>Natural Language Processing and Information Systems</source>
          , Springer Berlin Heidelberg, Berlin, Heidelberg,
          <year>2005</year>
          , pp.
          <fpage>227</fpage>
          -
          <lpage>238</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>M.</given-names>
            <surname>Le</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Roller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Papaxanthos</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kiela</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Nickel</surname>
          </string-name>
          ,
          <article-title>Inferring concept hierarchies from text corpora via hyperbolic embeddings</article-title>
          , in: A.
          <string-name>
            <surname>Korhonen</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Traum</surname>
          </string-name>
          , L. Màrquez (Eds.),
          <article-title>Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics</article-title>
          , Florence, Italy,
          <year>2019</year>
          , pp.
          <fpage>3231</fpage>
          -
          <lpage>3241</lpage>
          . doi:
          <volume>10</volume>
          . 18653/v1/
          <fpage>P19</fpage>
          - 1313.
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Ma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <article-title>A web-scale system for scientific knowledge exploration</article-title>
          , in: F. Liu, T. Solorio (Eds.),
          <source>Proceedings of ACL</source>
          <year>2018</year>
          ,
          <article-title>System Demonstrations, Association for Computational Linguistics</article-title>
          , Melbourne, Australia,
          <year>2018</year>
          , pp.
          <fpage>87</fpage>
          -
          <lpage>92</lpage>
          . doi:
          <volume>10</volume>
          .18653/v1/
          <fpage>P18</fpage>
          - 4015.
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <surname>OpenAlex</surname>
          </string-name>
          , Openalex:
          <article-title>End-to-end process for topic classification</article-title>
          ,
          <year>2024</year>
          . URL: https://docs. google.com/document/d/1bDopkhuGieQ4F8gGNj7sEc8WSE8mvLZS.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>G.</given-names>
            <surname>Wohlgenannt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Weichselbraun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Scharl</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Sabou</surname>
          </string-name>
          ,
          <article-title>Dynamic integration of multiple evidence sources for ontology learning</article-title>
          ,
          <source>Journal of Information and Data Management</source>
          <volume>3</volume>
          (
          <year>2012</year>
          )
          <fpage>243</fpage>
          -
          <lpage>254</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>J.</given-names>
            <surname>Mortensen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Musen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Noy</surname>
          </string-name>
          ,
          <article-title>Crowdsourcing the verification of relationships in biomedical ontologies</article-title>
          ,
          <source>Annual Symposium proceedings / AMIA Symposium. AMIA Symposium</source>
          <year>2013</year>
          (
          <year>2013</year>
          )
          <fpage>1020</fpage>
          -
          <lpage>9</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>B. P.</given-names>
            <surname>Allen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Stork</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Groth</surname>
          </string-name>
          ,
          <article-title>Knowledge engineering using large language models</article-title>
          ,
          <source>arXiv preprint arXiv:2310.00637</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>C.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lin</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Klein</surname>
          </string-name>
          ,
          <article-title>Constructing taxonomies from pretrained language models</article-title>
          ,
          <source>in: North American Chapter of the Association for Computational Linguistics</source>
          ,
          <year>2020</year>
          . URL: https://api.semanticscholar.org/CorpusID:233992529.
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24] Ansi/niso z39.
          <fpage>4</fpage>
          -
          <lpage>2021</lpage>
          , criteria for indexes,
          <year>2021</year>
          . URL: http://dx.doi.org/10.3789/ansi.niso.
          <year>z39</year>
          .
          <fpage>4</fpage>
          -
          <lpage>2021</lpage>
          . doi:
          <volume>10</volume>
          .3789/ansi.niso.
          <year>z39</year>
          .
          <fpage>4</fpage>
          -
          <lpage>2021</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>S.</given-names>
            <surname>Mukherjee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mitra</surname>
          </string-name>
          , G. Jawahar,
          <string-name>
            <given-names>S.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Palangi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Awadallah</surname>
          </string-name>
          , Orca:
          <article-title>Progressive learning from complex explanation traces of gpt-4</article-title>
          , arXiv preprint arXiv:
          <volume>2306</volume>
          .02707 (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>P.</given-names>
            <surname>Sharma</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. T.</given-names>
            <surname>Ash</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Misra</surname>
          </string-name>
          ,
          <article-title>The truth is in there: Improving reasoning in language models with layer-selective rank reduction</article-title>
          ,
          <source>arXiv preprint arXiv:2312.13558</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>P.</given-names>
            <surname>Yadav</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Tam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Choshen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. A.</given-names>
            <surname>Rafel</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Bansal</surname>
          </string-name>
          ,
          <article-title>Ties-merging: Resolving interference when merging models</article-title>
          ,
          <source>Advances in Neural Information Processing Systems</source>
          <volume>36</volume>
          (
          <year>2024</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <surname>T. de Bruin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Kober</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Tuyls</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Babuška</surname>
          </string-name>
          ,
          <article-title>Fine-tuning deep rl with gradient-free optimization</article-title>
          ,
          <source>IFAC-PapersOnLine</source>
          <volume>53</volume>
          (
          <year>2020</year>
          )
          <fpage>8049</fpage>
          -
          <lpage>8056</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>G.</given-names>
            <surname>Wang</surname>
          </string-name>
          , S. Cheng,
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Song</surname>
          </string-name>
          , Y. Liu, Openchat:
          <article-title>Advancing open-source language models with mixed-quality data</article-title>
          ,
          <source>arXiv preprint arXiv:2309.11235</source>
          (
          <year>2023</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>D.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Park</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Song</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Kim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Ahn</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Park</surname>
          </string-name>
          , G. Gim,
          <string-name>
            <given-names>M.</given-names>
            <surname>Cha</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Lee</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kim</surname>
          </string-name>
          , Solar
          <volume>10</volume>
          .7b:
          <article-title>Scaling large language models with simple yet efective depth up-scaling</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2312</volume>
          .
          <fpage>15166</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <article-title>AI@Meta, Llama 3 model card (</article-title>
          <year>2024</year>
          ). URL: https://github.com/meta-llama/llama3/blob/ main/MODEL_CARD.md.
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>