<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta>
      <issn pub-type="ppub">1613-0073</issn>
    </journal-meta>
    <article-meta>
      <title-group>
        <article-title>Few-Shot Approach for Relation Extraction Domain Adaptation using Large Language Models</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Vanni Zavarella</string-name>
          <email>vanni.zavarella@unica.it</email>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Juan Carlos Gamero-Salinas</string-name>
          <xref ref-type="aff" rid="aff2">2</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Sergio Consoli</string-name>
          <email>sergio.consoli@ec.europa.eu</email>
          <xref ref-type="aff" rid="aff1">1</xref>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Large Language Models, Knowledge Graphs, Few-shot Learning, Relation Extraction, Data Augmentation</string-name>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Department of Mathematics and Computer Science, University of Cagliari</institution>
          ,
          <addr-line>Cagliari</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>European Commission, Joint Research Centre (DG JRC)</institution>
          ,
          <addr-line>Ispra (VA)</addr-line>
          ,
          <country country="IT">Italy</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Institute of Data Science and Artificial Intelligence (DATAI), Universidad de Navarra</institution>
          ,
          <addr-line>Pamplona</addr-line>
          ,
          <country country="ES">Spain</country>
        </aff>
      </contrib-group>
      <abstract>
        <p>Knowledge graphs (KGs) have been successfully applied to the analysis of complex scientific and technological domains, with automatic KG generation methods typically building upon relation extraction models capturing fine-grained relations between domain entities in text. While these relations are fully applicable across scientific areas, existing models are trained on few domain-specific datasets such as SciERC and do not perform well on new target domains. In this paper, we experiment with leveraging incontext learning capabilities of Large Language Models to perform schema-constrained data annotation, collecting in-domain training instances for a Transformer-based relation extraction model deployed on titles and abstracts of research papers in the Architecture, Construction, Engineering and Operations (AECO) domain. By assessing the performance gain with respect to a baseline Deep Learning architecture trained on of-domain data, we show that by using a few-shot learning strategy with structured prompts and only minimal expert annotation the presented approach can potentially support domain adaptation of a science KG generation model.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>-</title>
      <p>CEUR
ceur-ws.org</p>
    </sec>
    <sec id="sec-2">
      <title>1. Introduction</title>
      <p>
        Knowledge graphs (KGs) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] have proved efective for representing research knowledge
discussed in scientific papers and patents across several diferent domains [
      </p>
      <sec id="sec-2-1">
        <title>2, 3, 4]. New generation</title>
        <p>“scientific KGs” have moved from representing purely bibliographic information of research
publications to support the construction of extensive networks of machine-readable information
about entities and relationships pertaining to a certain domain, enabling fine-grained semantic
queries over large scientific text collections such as: “retrieve all methods that are used for</p>
      </sec>
      <sec id="sec-2-2">
        <title>Indoor Air Remediation in the time range  ”.</title>
        <p>
          Therefore, they can support downstream analytical services like technology trend analysis.
For example, [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ] uses the statistics of relation triples of type &lt;Method;Used-for;Task&gt;
automatically extracted from paper abstracts to reconstruct historical trends of the top applications
of target methods such as “neural networks” in diferent areas like speech recognition and
computer vision.
(a)
(b)
        </p>
        <p>
          Methods for automatic generation of scientific KGs typically build upon training supervised
Relation Extraction (RE) models for capturing fine-grained relations between scientific entities
in text [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]. In the scholarly domain, the entity and relation specifications for this task are
defined by the SciERC initiative [
          <xref ref-type="bibr" rid="ref5">5</xref>
          ]1 and are fully applicable across scientific areas. As part of
an ongoing endeavour on innovation intelligence analytics for the Architecture, Engineering,
Construction and Operation (AECO) industry, we applied SciERC guidelines to annotate
domainspecific instances of scientific entities (e.g. Task, Method, Metrics) and their relations (e.g.
Usedfor,Evaluate-for ). In Figure1 we show a few sentence-level annotations of SciERC entity and
relation types respectively from an NLP paper abstract from the SCIERC dataset (a) and from
an AECO research paper annotated via the same tagging schema (b).
        </p>
        <p>
          One can see that the semantic schema is perfectly portable to the AECO domain. However,
state-of-the-art models for scholarly Relation Extraction have been trained on the SciERC dataset
comprising AI/ML articles and do not perform well on new target domains such as AECO[
          <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
          ].
Moreover, manually annotating training data for this domain is a costly and time-consuming
strategy that does not scale well.
        </p>
        <p>
          Therefore, we experiment on an empirical solution leveraging in-context learning capabilities
of Large Language Models (LLMs) [
          <xref ref-type="bibr" rid="ref10 ref11">10, 11</xref>
          ] to perform schema-constrained data annotation,
generating in-domain training instances for a baseline Relation Extraction architecture from only
        </p>
      </sec>
      <sec id="sec-2-3">
        <title>1https://nlp.cs.washington.edu/sciIE/annotation_guideline.pdf</title>
        <p>a few manually designed sentence annotation examples, together with explicit task instructions.
Then, we test diferent training configurations of the model on a small test set of titles and
abstracts of AECO research papers and compare the performance with a baseline trained on
out-of-domain data.</p>
        <p>Research on techniques for distilling knowledge from pre-trained LLMs to downstream
NLP tasks is currently highly active [12]. Prompt tuning approaches, that translate the target
downstream tasks to a masked language modeling problem have been applied to Relation
Classification [ 13, 14, 15]. Similarly to [15], we combine few-shot examples prompt with rich
schema information to elicit LLMs’ comprehension of the RE task. However, we explicitly
formulate the task as data annotation to the LLM.</p>
        <p>Finally, the presented approach is in line with the view from [12] that leveraging few-shot
learning capabilities of LLMs for optimizing local, lower-sized models is a more cost-efective
strategy than relying on direct use of LLMs for inference in a production setting, which typically
faces recurrent API usage costs or requires extensive high-end computational infrastructures
for fine-tuning.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>2. Data</title>
      <p>The source data used in this experiment comprise titles and abstracts from a large collection of
around 476k research articles in the AECO area published in the time range 2010-2023, retrieved
from the OpenAlex2 open scientific graph database [ 16] using a set of platform-specific topic
ifltering tags.</p>
      <p>We sampled a test set of around 50 abstracts, pre-processed and sentence split them using
Spacy’s English transformer pipeline en_core_web_trf-3.6.13 and finally had them independently
annotated by two domain experts using the Brat annotation tool [17], resulting in a total of 314
sentences, 448 entities, 132 relations instances. The inter-annotator positive specific agreement
on entity detection ([18]4) reached a mean F1 score of 0.73, indicating an overall satisfying
agreement between the human annotators, although some marginal ambiguity is encountered
for such a complex task. We publicly share the current version of the test dataset (called
SCIERC-AECO) in the github: https://github.com/zavavan/sperty/blob/main/datasets/scierc_
aec/scierc_aec_test.json and plan to release extended versions in the future.</p>
      <p>We used two random samples of respectively 3 and 10 sentences as example annotations for
the few-shot LLM prompts described below.</p>
    </sec>
    <sec id="sec-4">
      <title>3. Experimental Setups</title>
      <p>The full-stack Relation Extraction task consists of generating, for an input token sequence
 =  0, ...,   : a) a set of tuples  = &lt; (  , ..  ),   &gt; of typed token sub-sequences of  , with
0≤,  ≤ and   ∈   being a label belonging to the set   of entity labels; b) a set  of triples</p>
      <sec id="sec-4-1">
        <title>2https://docs.openalex.org/</title>
        <p>3https://github.com/explosion/spacy-models/releases/tag/en_core_web_trf-3.6.1
4This corresponds to classical Cohen κ inter-rater agreement, in tasks like NER where the number of negative cases
is undefined.
&lt; ℎ, ,   &gt; where ℎ,  ∈  are, respectively, the entity head and tail of the relation, and   ∈   is
a relation label.</p>
        <p>As baseline for the RE task we use SpERT (Span-based Entity and Relation Transformer) [19], a
span-based model for joint entity and relation extraction. SpERT is a relatively simple approach
using the pre-trained BERT for input token representation that classifies any arbitrary candidate
token span into entity types, filters non-entities ( None entity class) and finally classifies all pairs
of remaining entities into relations.</p>
        <p>By using only sentence-level context representations for sampling positive and negative
training examples, the architecture allows single-pass runs through BERT for each sentence,
resulting in significant speeding up of training. Despite this sentence-level RE simplification
though, SpERT significantly outperforms other joint entity/relation extraction models on SciERC
dataset, reaching up to 70.33% micro-average F1 on entity extraction and up to 50.84%
microaverage F1 on relation extraction (around 2.5% improvement on both tasks).</p>
        <p>We re-trained SpERT on SciERC training set (1861 sentences) using SciBERT (cased)
embeddings [20]. When tested over out-of-domain SCIERC-AECO data though, SpERT performance
degrades drastically. First row in Table 1 shows Micro-average F1 scores on SCIERC-AECO
for entity extraction (NER), relation detection without argument entity classification (RE) and
relation detection considering entity classification (RE_w/NEC), respectively.</p>
        <p>In order to test few-shot learning capability of LLMs for training data generation, we
experiment on schema-constrained instruction prompts sent to the Chat Completion endpoint
of the OpenAI gpt-3.5-turbo-0125 (ChatGPT) API [21]. The context length of the model is
approximately 4096 tokens.</p>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>4. Results and Discussion</title>
      <p>5Due to OpenAI API maximum request tokens limit, these are sent in batches of 10 sentences each.
6We trained for 20 epochs with 0.1 dropout rate. Experiments were run on NVIDIA A100-SXM4-40GB GPU device.
throughout the entire life cycle of the building have been reduced by 20.99%.” it generates an entity
not anchored in text(T1;Task;Carbon emissions reduction)7.</p>
      <p>Overall, the performance level is not outstanding across all configurations, considering that
the same model architecture is achieving a F1 measure of 2-3 factors higher when trained on
in-domain manually curated data (SciERC) of comparable size. This may be due to the model
degrading its generalization performance by learning from noisy labels [22], which is confirmed
by observing that the best results are obtained by adding ChatGPT generated labels to curated
out-of-domain SciERC labels. By considering only LLM-generated data, most configurations
slightly outperform the baseline for NER while only one does it for RE, indicating that this is an
harder task for LLM few-shot learning with respect to NER.</p>
      <p>Adding explicit Task definitions and increasing the number of few-shot examples both
consistently raise the performance with respect to all metrics, with the second finding seemingly
in contrast with what reported in [15].</p>
    </sec>
    <sec id="sec-6">
      <title>5. Conclusions</title>
      <p>This contribution presents our currently-ongoing work on the potentials of Large Language
Models (LLMs), specifically ChatGPT, for few-shot learning in the context of relation extraction
domain adaptation. In particular the study aimed to generate in-domain training data for a
Transformer-based relation extraction model within the Architecture, Construction,
Engineering, and Operations (AECO) domain by leveraging the in-context learning capabilities of LLMs.
The experiments involved using structured prompts and minimal expert annotation to collect
training instances from AECO research paper titles and abstracts.</p>
      <p>The results indicate that the quality of the LLM-generated annotations may not be suficient
to support domain customization of a RE model from ground up. However, when combined with
curated out-of-domain labels it can boost the performance on the new domain significantly.</p>
      <p>Overall, the research highlights the potential of using ChaptGPT for optimizing local,
lowersized models, which can be a more cost-efective strategy than relying on direct use of LLMs
for inference in production settings.</p>
      <p>Future work might include expanding the test set and conducting more extensive tests to
further validate the approach, also considering other domains than AECO. In addition, in the
future we plan to experiment with GPT-4 for data generation rather that ChatGPT, leveraging
its powerful capabilities to improve the quality of synthetic data, such as in the powerful LLaVA
multi-modal model [23].</p>
      <p>Object of further investigation and experiments will involve also the use of the latest advances
in open-source LLMs, such as by employing open-source models like Mistral-7B-OpenOrca8,
Nous Hermes Mixtral9, and Llama-3-70B10, to explore their potential in relation extraction
tasks. This could involve comparing the performance of these models directly with the current
7In a few other cases, annotation labels out of the schema are assigned to Entity and Relations.
8Mistral-7B-OpenOrca, https://huggingface.co/Open-Orca/Mistral-7B-OpenOrca
9Nous Hermes Mixtral, https://huggingface.co/NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO
10Llama-3-70B, https://huggingface.co/meta-llama/Meta-Llama-3-70B
approach, as well as exploring their capabilities in generating high-quality synthetic data for
ifne-tuning smaller models like SpERT.
R. Child, A. Ramesh, D. Ziegler, J. Wu, C. Winter, C. Hesse, M. Chen, E. Sigler, M. Litwin,
S. Gray, B. Chess, J. Clark, C. Berner, S. McCandlish, A. Radford, I. Sutskever, D. Amodei,
Language models are few-shot learners, in: H. Larochelle, M. Ranzato, R. Hadsell, M. Balcan,
H. Lin (Eds.), Advances in Neural Information Processing Systems, volume 33, Curran
Associates, Inc., 2020, pp. 1877–1901. URL: https://proceedings.neurips.cc/paper_files/
paper/2020/file/1457c0d6bfcb4967418bfb8ac142f64a-Paper.pdf.
[12] B. Ding, C. Qin, L. Liu, Y. K. Chia, S. Joty, B. Li, L. Bing, Is gpt-3 a good data annotator?,
2023. arXiv:2212.10450.
[13] J. Han, S. Zhao, B. Cheng, S. Ma, W. Lu, Generative prompt tuning for relation classification,
in: Y. Goldberg, Z. Kozareva, Y. Zhang (Eds.), Findings of the Association for Computational
Linguistics: EMNLP 2022, Association for Computational Linguistics, Abu Dhabi, United
Arab Emirates, 2022, pp. 3170–3185. URL: https://aclanthology.org/2022.findings-emnlp.
231. doi:10.18653/v1/2022.findings- emnlp.231.
[14] X. Chen, N. Zhang, X. Xie, S. Deng, Y. Yao, C. Tan, F. Huang, L. Si, H. Chen, Knowprompt:
Knowledge-aware prompt-tuning with synergistic optimization for relation extraction, in:
Proceedings of the ACM Web Conference 2022, WWW ’22, Association for Computing
Machinery, New York, NY, USA, 2022, p. 2778–2788. URL: https://doi.org/10.1145/3485447.
3511998. doi:10.1145/3485447.3511998.
[15] X. Xu, Y. Zhu, X. Wang, N. Zhang, How to unleash the power of large language models for
few-shot relation extraction?, 2023. arXiv:2305.01555.
[16] J. Priem, H. Piwowar, R. Orr, Openalex: A fully-open index of scholarly works, authors,
venues, institutions, and concepts, 2022. arXiv:2205.01833.
[17] P. Stenetorp, S. Pyysalo, G. Topić, T. Ohta, S. Ananiadou, J. Tsujii, brat: a web-based tool
for NLP-assisted text annotation, in: F. Segond (Ed.), Proceedings of the Demonstrations
at the 13th Conference of the European Chapter of the Association for Computational
Linguistics, Association for Computational Linguistics, Avignon, France, 2012, pp. 102–107.</p>
      <p>URL: https://aclanthology.org/E12-2021.
[18] G. Hripcsak, A. S. Rothschild, Agreement, the f-measure, and reliability in information
retrieval, Journal of the American Medical Informatics Association 12 (2005) 296–298.
URL: https://www.sciencedirect.com/science/article/pii/S1067502705000253. doi:https:
//doi.org/10.1197/jamia.M1733.
[19] M. Eberts, A. Ulges, Span-based joint entity and relation extraction with transformer
pre-training, in: ECAI 2020, IOS Press, 2020, pp. 2006–2013.
[20] I. Beltagy, K. Lo, A. Cohan, SciBERT: A pretrained language model for scientific text, in:
K. Inui, J. Jiang, V. Ng, X. Wan (Eds.), Proceedings of the 2019 Conference on Empirical
Methods in Natural Language Processing and the 9th International Joint Conference on
Natural Language Processing (EMNLP-IJCNLP), Association for Computational Linguistics,
Hong Kong, China, 2019, pp. 3615–3620. URL: https://aclanthology.org/D19-1371. doi:10.
18653/v1/D19- 1371.
[21] OpenAI, Text-davinci-003., https://platform.openai.com/docs/models/text-davinci-003,
2023.
[22] H. Song, M. Kim, D. Park, Y. Shin, J.-G. Lee, Learning from noisy labels with deep neural
networks: A survey, 2022. arXiv:2007.08199.
[23] H. Liu, C. Li, Q. Wu, Y. J. Lee, Visual instruction tuning, in: A. Oh,
T. Naumann, A. Globerson, K. Saenko, M. Hardt, S. Levine (Eds.), Advances
in Neural Information Processing Systems, volume 36, Curran Associates, Inc.,
2023, pp. 34892–34916. URL: https://proceedings.neurips.cc/paper_files/paper/2023/file/
6dcf277ea32ce3288914faf369fe6de0-Paper-Conference.pdf.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>S.</given-names>
            <surname>Ji</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Pan</surname>
          </string-name>
          , E. Cambria,
          <string-name>
            <given-names>P.</given-names>
            <surname>Marttinen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P. S.</given-names>
            <surname>Yu</surname>
          </string-name>
          ,
          <article-title>A survey on knowledge graphs: Representation, acquisition, and applications</article-title>
          ,
          <source>IEEE Transactions on Neural Networks and Learning Systems</source>
          <volume>33</volume>
          (
          <year>2022</year>
          )
          <fpage>494</fpage>
          -
          <lpage>514</lpage>
          . doi:
          <volume>10</volume>
          .1109/TNNLS.
          <year>2021</year>
          .
          <volume>3070843</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>D.</given-names>
            <surname>Dessì</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Osborne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. Reforgiato</given-names>
            <surname>Recupero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Buscaldi</surname>
          </string-name>
          , E. Motta,
          <article-title>Generating knowledge graphs by employing natural language processing and machine learning techniques within the scholarly domain</article-title>
          ,
          <source>Future Generation Computer Systems</source>
          <volume>116</volume>
          (
          <year>2021</year>
          )
          <fpage>253</fpage>
          -
          <lpage>264</lpage>
          . URL: https://www.sciencedirect.com/science/article/pii/S0167739X2033003X. doi:https://doi. org/10.1016/j.future.
          <year>2020</year>
          .
          <volume>10</volume>
          .026.
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>L.</given-names>
            <surname>Siddharth</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. T. M.</given-names>
            <surname>Blessing</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K. L.</given-names>
            <surname>Wood</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Luo</surname>
          </string-name>
          ,
          <article-title>Engineering Knowledge Graph From Patent Database</article-title>
          ,
          <source>Journal of Computing and Information Science in Engineering</source>
          <volume>22</volume>
          (
          <year>2021</year>
          )
          <article-title>021008</article-title>
          . URL: https://doi.org/10.1115/1.4052293. doi:
          <volume>10</volume>
          .1115/1.4052293.
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Xiao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Xiao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Thürer</surname>
          </string-name>
          ,
          <article-title>A patent recommendation method based on kg representation learning</article-title>
          ,
          <source>Engineering Applications of Artificial Intelligence</source>
          <volume>126</volume>
          (
          <year>2023</year>
          ). doi:
          <volume>10</volume>
          .1016/j. engappai.
          <year>2023</year>
          .
          <volume>106722</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Luan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>He</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Ostendorf</surname>
          </string-name>
          ,
          <string-name>
            <surname>H.</surname>
          </string-name>
          <article-title>Hajishirzi, Multi-task identification of entities, relations, and coreferencefor scientific knowledge graph construction</article-title>
          ,
          <source>in: Proc. Conf. Empirical Methods Natural Language Process</source>
          .
          <source>(EMNLP)</source>
          ,
          <year>2018</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>X.</given-names>
            <surname>Zhao</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Deng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Zhang</surname>
          </string-name>
          , H. Cheng, W. Lam,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <article-title>A comprehensive survey on deep learning for relation extraction: Recent advances</article-title>
          and new frontiers,
          <year>2023</year>
          . arXiv:
          <fpage>2306</fpage>
          .
          <year>02051</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>M.</given-names>
            <surname>Han</surname>
          </string-name>
          ,
          <string-name>
            <surname>J</surname>
          </string-name>
          . Park, I. Kim,
          <string-name>
            <given-names>H.</given-names>
            <surname>Yi</surname>
          </string-name>
          ,
          <article-title>A microalgae photobioreactor system for indoor air remediation: Empirical examination of the co2 absorption performance of spirulina maxima in a nahco3-reduced medium</article-title>
          ,
          <source>Applied Sciences</source>
          <volume>13</volume>
          (
          <year>2023</year>
          ). URL: https://www.mdpi.com/ 2076-3417/13/24/12991. doi:
          <volume>10</volume>
          .3390/app132412991.
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>D.</given-names>
            <surname>Dessí</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Osborne</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. R.</given-names>
            <surname>Recupero</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Buscaldi</surname>
          </string-name>
          , E. Motta,
          <article-title>Scicero: A deep learning and nlp approach for generating scientific knowledge graphs in the computer science domain</article-title>
          ,
          <source>Knowledge-Based Systems</source>
          <volume>258</volume>
          (
          <year>2022</year>
          ). URL: https://oro.open.ac.uk/85472/. doi:
          <volume>10</volume>
          .1016/j. knosys.
          <year>2022</year>
          .
          <volume>109945</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Z.</given-names>
            <surname>Zhong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <article-title>A frustratingly easy approach for entity and relation extraction</article-title>
          , in: K.
          <string-name>
            <surname>Toutanova</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Rumshisky</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Zettlemoyer</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Hakkani-Tur</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <string-name>
            <surname>Beltagy</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Bethard</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Cotterell</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Chakraborty</surname>
          </string-name>
          , Y. Zhou (Eds.),
          <source>Proceedings of the</source>
          <year>2021</year>
          <article-title>Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Association for Computational Linguistics</article-title>
          , Online,
          <year>2021</year>
          , pp.
          <fpage>50</fpage>
          -
          <lpage>61</lpage>
          . URL: https://aclanthology.org/
          <year>2021</year>
          .naacl-main.5. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2021</year>
          .naacl- main.5.
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>Q.</given-names>
            <surname>Dong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Dai</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Zheng</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Wu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Chang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>X.</given-names>
            <surname>Sun</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Li</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Sui</surname>
          </string-name>
          ,
          <article-title>A survey on in-context learning</article-title>
          ,
          <year>2023</year>
          . arXiv:
          <volume>2301</volume>
          .
          <fpage>00234</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>T.</given-names>
            <surname>Brown</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Mann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Ryder</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Subbiah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. D.</given-names>
            <surname>Kaplan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Dhariwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Neelakantan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Shyam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>G.</given-names>
            <surname>Sastry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Askell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Agarwal</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Herbert-Voss</surname>
          </string-name>
          , G. Krueger, T. Henighan,
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>