<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>SynSem-Align (Demo): Ontology-Driven KG Extraction via Syntactic Candidate Mining and Paraphrase-Based Equivalence Filtering</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Rikuto Sasaki</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Masahito Yasui</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kazuhiro Takeuchi</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>Osaka Electro-Communication University</institution>
        </aff>
      </contrib-group>
      <abstract>
        <p>While Large Language Models (LLMs) are powerful for information extraction, the reliability of their output remains a challenge, making human supervision essential. We introduce SynSem-Align, a support tool where LLMs and humans collaborate on knowledge extraction. Our approach integrates three core components: (1) Ontology-Driven filtering to suggest relevant extraction patterns, (2) Syntactic Candidate Mining to precisely identify knowledge candidates using a CKY-based approach over dependency structures, and (3) ParaphraseBased Equivalence Filtering using an LLM for semantic validation. This integrated workflow enables users to transparently and reliably construct knowledge graphs, demonstrating a practical path towards verifiable knowledge extraction that balances automation with human oversight.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Knowledge Graph</kwd>
        <kwd>Ontology</kwd>
        <kwd>Natural language processing</kwd>
        <kwd>Large Language Model</kwd>
        <kwd>Semantic analysis</kwd>
        <kwd>Information extraction</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        While Large Language Models (LLMs) have significantly advanced automated knowledge extraction,
ensuring the reliability and verifiability of their outputs remains a key challenge [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. End-to-end
extraction methods that rely solely on LLMs often lack transparency, making it dificult for human
experts to verify results or correct errors. Because the internal reasoning of these models is not directly
observable, guaranteeing its faithfulness remains challenging [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>In this paper, we present SynSem-Align, a tool for verifiable, human-supervised knowledge extraction.
SynSem-Align externalizes the extraction logic as explicit, selectable patterns, rather than relying on
implicit, model-internal processing. In contrast to black-box approaches with LLMs that attempt to
generate triples directly from prompts embedding both the source text and ontology constraints, our
tool first enumerates multiple candidate triples and then refines them using ontology-based filtering
and paraphrase-based equivalence checks. This combination ensures that extraction is not only guided
by linguistic constraints but also accompanied by reasoning steps that can be inspected and corrected.
A distinctive aspect of SynSem-Align is that it explicitly acknowledges inherent ambiguities: even after
refinement, alternative knowledge representations may remain. Instead of suppressing this diversity,
SynSem-Align provides mechanisms to expose it to human judgment. Through an interactive interface
for visual verification, users can examine how textual spans map onto candidate triples and adjudicate
among competing interpretations. This design allows knowledge graphs to be constructed in a manner
that is transparent, auditable, and domain-adaptable.</p>
      <p>To complement these methodological contributions, we provide a demonstration that walks through
SynSem-Align’s workflow step by step. To ensure reproducibility and foster further exploration, the
source code is publicly available on GitHub1.</p>
    </sec>
    <sec id="sec-2">
      <title>2. One Text Span Admits Many Candidate Knowledge Graph Triples</title>
      <p>A fundamental task in knowledge extraction is to make explicit, for a specific span of text, what
knowledge can justifiably be extracted and why. In our view, the core dificulty is not a simple opposition
between syntax and semantics but a residual gap between the results of syntactic and semantic analyses
and the knowledge graph that we finally construct. Syntactic and semantic analyses constrain the space
of possible meanings, yet the set of valid interpretations remains diverse, and that diversity is often
domain dependent even for expressions that share the same surface structure.</p>
      <p>Consider the metonymic phrase ‘Land of Smiles’. The ‘A of B’ construction typically ofers  and 
as candidates for relations in the knowledge graph, for example attribution, origin, or possession; in
this instance a natural interpretation is that the residents in  frequently smile (). This illustrates that
even text spans that share the same syntactic ‘A of B’ structure can yield diverse semantic relationships
between their terms. Furthermore, even after applying both syntactic and semantic constraints, the
mapping from the analyzed text to a knowledge graph may still be indeterminate. If we admit a
latent variable  that does not appear in the span, the partial expression ‘A of B’ can license multiple
candidate relations, for example  (, ), (, ), (, ), and others. This indicates that, despite
the restrictions provided by syntax and semantics, a substantial set of candidate knowledge graphs
remains. Since the appropriate choice can vary by domain or field, tools are needed that make the
mapping from text to knowledge graph relations explicit and verifiable on the basis of consistent
reasoning.</p>
      <p>End to end LLMs may learn such correspondences implicitly, but their black box nature prevents us
from verifying how a particular triple was licensed by a particular piece of text, and it prevents users
from trusting or correcting the outcome. We therefore emphasize transparency and controllability.
Users should see which part of the text supports which candidate relation, and what reasoning connects
them. In practice, we operationalize this with paraphrase based validation anchored at the span: if a
syntactic candidate can be rephrased into a natural and semantically equivalent sentence aligned with
the same text span, it is retained; otherwise, it is discarded. This procedure enumerates the diversity of
interpretations that remain after syntax and semantics have constrained the space, and it justifies each
retained interpretation in a way that can be inspected, trusted, and corrected by humans.</p>
    </sec>
    <sec id="sec-3">
      <title>3. Related Work</title>
      <p>
        The dominant approach for knowledge extraction currently leverages the zero-shot or few-shot
capabilities of Large Language Models (LLMs) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]. While powerful, these end-to-end methods sufer from a
lack of transparency: the reasoning process by which an LLM maps sentence structures to ontological
relations is inaccessible to users. Several works have attempted to expose the model’s intermediate
reasoning, but the reliability of such explanations remains uncertain [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>
        Our work takes a diferent stance by grounding the extraction process in a pre-constructed pattern
base governed by humans. This pattern base is semi-automatically derived from text-to-graph datasets
such as Text2KGBench [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], which provide systematically aligned text–triple pairs and thus a practical
foundation for reusable extraction rules. In SynSem-Align, the LLM is not used for end-to-end triple
generation. Its role is narrowly confined to auxiliary tasks such as paraphrase-based validation. As a
result, the rationale for each extracted triple is not hidden within an opaque model but is anchored in
explicit patterns that users can inspect and control. This ensures that the extraction process remains
transparent and verifiable. This line of research extends our earlier work on domain-specific knowledge
extraction [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] and its enhancement with generative models [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ].
      </p>
      <p>
        SynSem-Align enhances transparency and makes explicit the assumptions underlying LLM reasoning
by explicitly controlling syntactic and semantic judgments, thereby supporting users in making final
decisions regarding knowledge extraction. This stance stands in contrast to the Auto-KG Agent [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ],
which has been proposed as a multi tool framework where LLMs act as agents to invoke relation
extraction systems such as REBEL [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] and KnowGL [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ], integrate the extracted triples, and re rank them.
REBEL is a seq2seq model based on BART-large that linearizes entity mentions, types, and relations
for end to end extraction, while KnowGL combines knowledge generation, fact ranking, and Wikidata
linking through fine tuned language models. By orchestrating these components via an LLM agent,
Auto-KG Agent pursues autonomy and aims to improve triple extraction accuracy, particularly for
complex sentences or those involving negation, while minimizing direct human involvement. Whereas
Auto-KG Agent seeks to reduce the human role in the extraction process, SynSem-Align establishes an
alternative paradigm in which eficiency is carefully balanced with verifiability, ensuring that extracted
knowledge remains interpretable and under explicit human control.
      </p>
    </sec>
    <sec id="sec-4">
      <title>4. Transparent Human Supervised Workflow of SynSem-Align</title>
      <p>SynSem-Align structures the extraction process into explicit, inspectable steps. The workflow
systematically enumerates syntactic candidates, prunes implausible ones through ontology based and paraphrase
based validation, and reserves the final decision for the human user. This staged design ensures that
the diversity of possible interpretations is preserved, implausible ones are pruned, and the outcome
remains transparent and auditable.</p>
      <sec id="sec-4-1">
        <title>4.1. Enumerating a Broad Set of Syntactic Candidates</title>
        <p>
          The process begins with dependency parsing, which identifies grammatical relations and decomposes
the syntactic units(Figure 1). This analysis enables the system to process complex constructions such as
coordination [
          <xref ref-type="bibr" rid="ref10 ref9">9, 10</xref>
          ], ensuring that conjunctive phrases are segmented into appropriate substructures.
        </p>
        <p>The resulting structure feeds into a CKY-based charting algorithm [11], which combines sentence
fragments and systematically matches them against the pre-constructed pattern base (Figure 2). The
pattern base consists of reusable templates such as ‘[X1] is [Y1]’ or ‘[X1] of [Y1],’ which explicitly link
surface syntax to potential semantic relations. Through this process, the system enumerates all triples
that the surface syntax can license, anchored to explicit and reusable patterns rather than opaque model
inferences.</p>
        <p>The outcome is a broad but systematically organized set of syntactic candidates, which inherently
inclN映uO画dU「eNs　 bとなoりth　　　p　l　aのu　s　i　b　leトaトnロ」d　　s　pのu　r　i　o　　u　s監i督n　t　e　r　pと　r　e　t　a　tio脚n本s　.　　A　tはt　h　i　s　　sta宮g崎e　,　　th　e駿g　o　a　　l　iでs exhPRaOuPsNtive ePnROuPmNeratAiUoXn、,</p>
        <p>NOUN ADP PROPN ADP NOUN ADP NOUN ADP
leavmionvige, thneexttadosokr of dNiOscardingToitomro plausNiOble canddiirdecatotres toTOsubsescqreuenewnritterrefineHmA ent staMgiyaezsak.i Hayao DE
映画 「となり の トトロ」 の
NOUN NOUN ADP PROPN ADP
movie next door NO Totoro NO
(Tonari no Totoro, instance, movie)　　　　
　　　(Hayao Miyazaki, director ,Tonari no Totoro )　
監督
NOUN
director
と
ADP
TO
脚本 は
NOUN ADP
screen writer HA
宮崎
PROPN
Miyazaki
駿
PROPN
Hayao
で</p>
        <p>AUX</p>
        <p>DE
&amp;　　
(Hayao Miyazaki, screen writer,Tonari no Totoro )　 　</p>
      </sec>
      <sec id="sec-4-2">
        <title>4.2. Refining Candidates through Ontology and Paraphrase Validation</title>
        <p>Because the syntactic candidate set is inherently diverse, it inevitably contains both plausible and
spurious interpretations. This diversity is deliberately preserved, since subsequent refinement ensures
that only semantically valid interpretations remain.</p>
        <p>First, ontology based type and relation constraints rule out triples inconsistent with domain knowledge
(Figure 3). For example, the system maps ‘My Neighbor Totoro’ to the concept film and ‘director’ to
the relation director, discarding implausible triples such as ‘film directed person’.
映画「となりのトトロ」の　　　　 監督と　　　　 　 脚本は　　　　 宮崎駿で、 
movie ’Tonari no Totoro’ -NO director-TO screen writer-HA Miyazaki Hayao-DE</p>
      </sec>
      <sec id="sec-4-3">
        <title>4.3. Finalizing Knowledge through Human Guided Selection</title>
        <p>Even after syntactic enumeration and refinement through ontology and paraphrase validation, multiple
semantically valid candidates may remain. These reflect residual ambiguity that cannot be resolved
automatically, and thus the decisive step is delegated to the human user.</p>
        <p>The remaining candidates represented as abstract patterns such as ‘[X1] is [Y1]’ or ‘[X1] of [Y1]’ are
presented in an explicit and inspectable form. The user visually inspects these alternatives and selects
the pattern that best captures the intended meaning of the input sentence.</p>
        <p>This explicit human decision directly determines the extracted knowledge and ensures that the
reasoning process remains auditable: users can see which syntactic candidates were generated, which
were excluded, and why the final pattern was chosen. The division of responsibilities is therefore clear:
the system enumerates and refines candidates, the LLM provides only bounded auxiliary validation,</p>
        <sec id="sec-4-3-1">
          <title>Purpose: Rigorously verify the following two points about the parallel elements in the target sentence and output only “True” or “False” in JSON format.</title>
        </sec>
        <sec id="sec-4-3-2">
          <title>Input:</title>
          <p>- movie ’tonari no Totoro’ -NO director-TO screen writer-HA Miyazaki Hayao-DE
- Parallel elements: “director”, “screen writer” (e.g., "A", "B", "C")
Judgment Criteria:
1. Similarity Check:
- The head (central word) of each parallel element must be the same part of speech
(e.g., all nouns or all verbs).
- Phrase structures should be aligned as much as possible</p>
          <p>(e.g., all “noun + particle”).</p>
        </sec>
        <sec id="sec-4-3-3">
          <title>2. Commutability Check:</title>
          <p>- Swapping the order of the parallel elements must still produce
a natural Japanese sentence.
- Swapping only the parallel elements should not significantly change
the overall meaning of the sentence (the main roles/semantic structure are
preserved).</p>
        </sec>
        <sec id="sec-4-3-4">
          <title>Decision Rule:</title>
          <p>- Output “True” only if both criteria are satisfied,</p>
          <p>output “False” if either criterion is not met.
- The output must be **only** the following JSON format.</p>
          <p>Do **not** include any explanations or comments.
and the final disambiguation rests with the human expert. In this way, SynSem-Align guarantees that
knowledge extraction remains transparent, controllable, and accountable, with interpretive authority
explicitly assigned to the user.</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>5. Conclusion</title>
      <p>We have presented SynSem-Align, a human supervised knowledge extraction tool. By efectively
combining an ontology driven filter, a transparent syntactic matching engine based on dependency parsing
and CKY, and LLM assisted semantic validation, our tool ofers a transparent and verifiable workflow
for supervised extraction.The explicit, pattern based logic, controlled by the user and verified through
semantic checks, provides a robust framework for building reliable knowledge graphs, successfully
balancing automation with essential human oversight.</p>
    </sec>
    <sec id="sec-6">
      <title>Declaration on Generative AI</title>
      <p>ChatGPT (OpenAI, 2025) was used only for English language checks (rephrasing, grammar, and style).
All scientific content was created by the authors, who take full responsibility for the final manuscript.</p>
    </sec>
    <sec id="sec-7">
      <title>Acknowledgment</title>
      <p>This research was partially supported by JSPS KAKENHI Grant Number JP23K28152.
Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers),
Association for Computational Linguistics, Minneapolis, Minnesota, 2019, pp. 3394–3403. URL:
https://aclanthology.org/N19-1343/. doi:10.18653/v1/N19-1343.
[11] T. KASAMI, An eficient recognition and syntax analysis algorithm for context-free languages,
Science Report, Air Force Cambridge Research Laboratory (1965).</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <surname>T. B. Brown</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Mann</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Ryder</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Subbiah</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Kaplan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Dhariwal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Neelakantan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          <string-name>
            <surname>Shyam</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          <string-name>
            <surname>Sastry</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Askell</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Agarwal</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Herbert-Voss</surname>
            , G. Krueger,
            <given-names>T.</given-names>
          </string-name>
          <string-name>
            <surname>Henighan</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          <string-name>
            <surname>Child</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Ramesh</surname>
            ,
            <given-names>D. M.</given-names>
          </string-name>
          <string-name>
            <surname>Ziegler</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Winter</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Hesse</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Chen</surname>
            , E. Sigler,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Litwin</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Gray</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Chess</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Clark</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Berner</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>McCandlish</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          <string-name>
            <surname>Radford</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          <string-name>
            <surname>Sutskever</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          <string-name>
            <surname>Amodei</surname>
          </string-name>
          ,
          <article-title>Language models are few-shot learners</article-title>
          ,
          <source>in: Advances in Neural Information Processing Systems</source>
          <volume>33</volume>
          (NeurIPS
          <year>2020</year>
          ),
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>T.</given-names>
            <surname>Korbak</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Balesni</surname>
          </string-name>
          , E. Barnes,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Bengio</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Benton</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Bloom</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Chen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Cooney</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dafoe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Dragan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Emmons</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Evans</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Farhi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Greenblatt</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Hendrycks</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hobbhahn</surname>
          </string-name>
          , E. Hubinger,
          <string-name>
            <given-names>G.</given-names>
            <surname>Irving</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Jenner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Kokotajlo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Krakovna</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Legg</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Lindner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Luan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Mądry</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Michael</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Nanda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Orr</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Pachocki</surname>
          </string-name>
          , E. Perez,
          <string-name>
            <given-names>M.</given-names>
            <surname>Phuong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Roger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Saxe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Shlegeris</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Soto</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Steinberger</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Zaremba</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Baker</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Shah</surname>
          </string-name>
          ,
          <string-name>
            <given-names>V.</given-names>
            <surname>Mikulik</surname>
          </string-name>
          ,
          <article-title>Chain of thought monitorability: A new and fragile opportunity for ai safety</article-title>
          ,
          <year>2025</year>
          . URL: https://arxiv.org/abs/2507.11473. arXiv:
          <volume>2507</volume>
          .
          <fpage>11473</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>N.</given-names>
            <surname>Mihindukulasooriya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Tiwari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. F.</given-names>
            <surname>Enguix</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lata</surname>
          </string-name>
          ,
          <article-title>Text2kgbench: A benchmark for ontologydriven knowledge graph generation from text</article-title>
          ,
          <source>in: Proceedings of the 22nd International Semantic Web Conference (ISWC</source>
          <year>2023</year>
          ),
          <year>2023</year>
          , pp.
          <fpage>247</fpage>
          -
          <lpage>265</lpage>
          . doi:
          <volume>10</volume>
          .1007/978-3-
          <fpage>031</fpage>
          -47243-5_
          <fpage>14</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>M.</given-names>
            <surname>YASUI</surname>
          </string-name>
          ,
          <string-name>
            <surname>T. KITAJIMA</surname>
          </string-name>
          , H. TANIDA,
          <string-name>
            <given-names>K.</given-names>
            <surname>MIYOSHI</surname>
          </string-name>
          ,
          <string-name>
            <surname>S. OBA</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>TAKEUCHI</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. KOZAKI</surname>
          </string-name>
          ,
          <article-title>Preprocessing for extracting knowledge from job and skill definition sentences with a simplified sentence pattern description</article-title>
          ,
          <source>Proceedings of the Annual Conference of JSAI JSAI2025</source>
          (
          <year>2025</year>
          )
          <fpage>2Win524</fpage>
          -
          <lpage>2Win524</lpage>
          . doi:
          <volume>10</volume>
          .11517/pjsai.
          <source>JSAI2025</source>
          .
          <volume>0</volume>
          _
          <fpage>2Win524</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>YASUI</surname>
          </string-name>
          ,
          <string-name>
            <surname>K. TAKEUCHI</surname>
          </string-name>
          ,
          <article-title>Enhancing a tool for extracting knowledge graphs from text through the utilization of generative language models</article-title>
          ,
          <source>JSAI Technical Report, Type 2 SIG</source>
          <year>2024</year>
          (
          <year>2024</year>
          )
          <article-title>04</article-title>
          . doi:
          <volume>10</volume>
          .11517/jsaisigtwo.
          <year>2024</year>
          .SWO-
          <volume>063</volume>
          _
          <fpage>04</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>A.</given-names>
            <surname>Ananya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Tiwari</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Mihindukulasooriya</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Soru</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Xu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Moussallem</surname>
          </string-name>
          ,
          <article-title>Towards harnessing large language models as autonomous agents for semantic triple extraction from unstructured text</article-title>
          ,
          <source>in: Extended Semantic Web Conference</source>
          ,
          <year>2024</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <surname>P.-L. Huguet</surname>
            <given-names>Cabot</given-names>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Navigli</surname>
          </string-name>
          , REBEL:
          <article-title>Relation extraction by end-to-end language generation</article-title>
          , in: M.
          <article-title>-</article-title>
          <string-name>
            <surname>F. Moens</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Specia</surname>
          </string-name>
          , S. W.-t. Yih (Eds.),
          <source>Findings of the Association for Computational Linguistics: EMNLP</source>
          <year>2021</year>
          ,
          <article-title>Association for Computational Linguistics</article-title>
          , Punta Cana, Dominican Republic,
          <year>2021</year>
          , pp.
          <fpage>2370</fpage>
          -
          <lpage>2381</lpage>
          . URL: https://aclanthology.org/
          <year>2021</year>
          .findings-emnlp.
          <volume>204</volume>
          /. doi:
          <volume>10</volume>
          . 18653/v1/
          <year>2021</year>
          .findings-emnlp.
          <volume>204</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>G.</given-names>
            <surname>Rossiello</surname>
          </string-name>
          ,
          <string-name>
            <surname>M. F. M. Chowdhury</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          <string-name>
            <surname>Mihindukulasooriya</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          <string-name>
            <surname>Cornec</surname>
            ,
            <given-names>A. M.</given-names>
          </string-name>
          <string-name>
            <surname>Gliozzo</surname>
          </string-name>
          ,
          <article-title>Knowgl: Knowledge generation and linking from text</article-title>
          , in: AAAI, AAAI Press,
          <year>2023</year>
          , pp.
          <fpage>16476</fpage>
          -
          <lpage>16478</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>Y.</given-names>
            <surname>Sawada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Wada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Shibahara</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Teranishi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kondo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Shindo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Watanabe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Matsumoto</surname>
          </string-name>
          ,
          <article-title>Coordination boundary identification without labeled data for compound terms disambiguation</article-title>
          ,
          <source>in: Proceedings of the 28th International Conference on Computational Linguistics (COLING</source>
          <year>2020</year>
          ),
          <source>International Committee on Computational Linguistics</source>
          , Barcelona,
          <source>Spain (Online)</source>
          ,
          <year>2020</year>
          , pp.
          <fpage>3043</fpage>
          -
          <lpage>3049</lpage>
          . URL: https://aclanthology.org/
          <year>2020</year>
          .coling-main.
          <volume>271</volume>
          /. doi:
          <volume>10</volume>
          .18653/v1/
          <year>2020</year>
          . coling-main.
          <volume>271</volume>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <given-names>H.</given-names>
            <surname>Teranishi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Shindo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Matsumoto</surname>
          </string-name>
          ,
          <article-title>Decomposed local models for coordinate structure parsing</article-title>
          ,
          <source>in: Proceedings of the 2019 Conference of the North American Chapter of the Association for</source>
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>