Ontology-grounded Automatic Knowledge Graph Construction by LLM under Wikidata schema Xiaohan Feng1 , Xixin Wu1 and Helen Meng1,* 1 Department of System Engineering and Engineering Management, Chinese University of Hong Kong Abstract We propose an ontology-grounded approach to Knowledge Graph (KG) construction using Large Lan- guage Models (LLMs) on a knowledge base. An ontology is authored by generating Competency Questions (CQ) on knowledge base to discover knowledge scope, extracting relations from CQs, and attempt to replace equivalent relations by their counterpart in Wikidata. To ensure consistency and interpretability in the resulting KG, we ground generation of KG with the authored ontology based on extracted relations. Evaluation on benchmark datasets demonstrates competitive performance in knowledge graph construc- tion task. Our work presents a promising direction for scalable KG construction pipeline with minimal human intervention, that yields high quality and human-interpretable KGs, which are interoperable with Wikidata semantics for potential knowledge base expansion. Keywords Knowledge Graph, Relation Extraction, Large Language Model, Wikidata, Interpretable AI 1. Introduction Knowledge Graphs (KGs) are structured representations of information that capture entities and their relationships in a graph format. By organizing knowledge in a machine-readable way, KGs enable a wide range of intelligent applications, such as semantic search, question answering, recommendation systems, and decision support [1]. The ability to construct high- quality, comprehensive KGs is thus critical for harnessing the power of these technologies across various domains. Traditionally, the process of constructing KGs has relied heavily on manual effort by domain experts to define the relevant entities and relationships, populate the graph with valid facts, and ensure logical consistency [2]. However, this manual curation approach is time-consuming, expensive, and difficult to scale to large, evolving domains. There is a strong need for (semi- )automatic methods that can aid the KG construction process by extracting structured knowledge from unstructured data sources such as text. Recent years have seen growing interest in leveraging Large Language Models (LLMs) for various knowledge capture and reasoning tasks [3]. Pre-trained on vast amounts of text data, LLMs can generate fluent natural language and have been shown to memorize and recall factual knowledge [4], [5]. However, directly applying LLMs to KG construction still faces several challenges. First, LLMs may generate inconsistent or redundant facts due to the lack HI-AI@KDD, Human-Interpretable AI Workshop at the KDD 2024, 26π‘‘β„Ž of August 2024, Barcelona, Spain * Corresponding author. $ xhfeng@se.cuhk.edu.hk (X. Feng); wuxx@se.cuhk.edu.hk (X. Wu); hmmeng@se.cuhk.edu.hk (H. Meng) Β© 2024 Copyright for this paper by its authors. Use permitted under Creative Commons License Attribution 4.0 International (CC BY 4.0). CEUR ceur-ws.org Workshop ISSN 1613-0073 Proceedings of an explicit, unified schema [6]. Second, the generated KGs may be incomplete or biased towards the knowledge present in the LLM’s training data, which may not fully cover the target domain, especially for proprietary documents not included in pre-training set. Finally, it can be challenging to integrate LLM-generated KGs with existing knowledge bases due to misalignment with standard ontologies. In this work, we propose a novel approach that harnesses the reasoning power of LLMs and the structured schema of Wikidata to construct high-quality KGs for proprietary knowledge domains. Our approach begins by discovering the scope of knowledge through the generation of Competency Questions (CQ) and answers from unstructured documents. We then summarize the relations and prop- erties from these QA pairs into an ontology, matching candidate properties against those defined in Wikidata and extending the schema as needed. Finally, we use the resulting on- Figure 1: Flowchart of proposed approach. tology to ground the transformation of CQ- answer pairs into a structured KG. By incorporating the Wikidata schema into our pipeline and grounding generation of KG on the same ontology, we aim to reduce redundancy, leverage the implicit knowledge captured during LLM pretraining while improving interpretability, and ensure interoperability with public knowledge bases. The generated KGs could be parsed with RDF parsers and used in downstream applications, or audited for correctness. The main contributions of this work are as follows: 1. We propose a novel ontology-grounded approach to LLM-based KG construction that leverages ontology based on Wikidata schema to guide the extraction and integration of knowledge from unstructured text. 2. We introduce a pipeline that combines competency question generation, ontology align- ment, and KG grounding to systematically construct high-quality KGs that are consistent, complete, and interoperable with existing knowledge bases. 3. We demonstrate the effectiveness of our approach through experiments on benchmark datasets, showing improvements in KG quality compared to traditional methods alongside with interpretability and utility of generated KGs. 2. Literature Review Knowledge graph construction has been an active area of research in recent years, with a wide range of approaches proposed for extracting structured knowledge from unstructured data sources [2]. Early methods relied heavily on rule-based systems and hand-crafted features to identify entities and relations in text [7]. With the advent of deep learning, neural network- based approaches have become increasingly popular, enabling more flexible and scalable KG construction [8]. One prominent line of work focuses on using distant supervision to automatically generate training data for relation extraction [9]. These methods assume that if two entities are mentioned together in a sentence and also appear in a knowledge base as subject and object of a relation, then that sentence is likely to express the relation. While distant supervision has been shown to be effective at scale, it often suffers from noise and incomplete coverage. Another important direction is the development of unsupervised and semi-supervised meth- ods for KG construction [10]. These approaches aim to reduce the reliance on large amounts of labeled data by leveraging techniques such as bootstrapping, graph-based inference, and representation learning. However, they often struggle with consistency and quality control issues. More recently, there has been growing interest in using large language models for KG construction [11], [12], [13]. These methods take advantage of the vast knowledge cap- tured in pretrained Language Models (LM) to generate KG triples through prompt engineering and fine-tuning. While promising, these LM-based approaches only produces triplets without canocalization, which makes portability and interoperability difficult. Additionally, some meth- ods rely on vector-based similarity measures to deduce relationships between entities in KG, which yields good performance but falls short in interpretability [14]. As mentiond in Introduction, despite the significant progress in KG construction and LLM applications, performance, interpretability, coverage of proprietary documents, and interaction with other knowledge base remain issues. Our pipeline address these by grouding KG generation on ontology based on Wikidata schema, which ensures that output KG is human-readable and makes integrating with Wikidata or other KG easier; In the experiments below we show that these benefits can be also achieved on private documents with decent performance. 3. Method: Ontology-grounded KG Construction Our proposed approach for ontology-grounded KG construction using LLMs consists of four main stages: 1) Competency Question Generation, 2) Relation Extraction and Ontology Matching, 3) Ontology Formatting, and 4) KG Construction. Figure 1 provides an overview of the pipeline. 3.1. Competency Question (CQ)-Answer Generation The first step in our pipeline is to generate a set of competency questions (CQs) and answers that capture the key information needs of the target domain. We employ an LLM to generate CQs based on the input documents. The LLM is provided with a set of instructions and examples to guide the generation process, encouraging the creation of well-formed, relevant questions that can be answered using the given documents. This step helps to scope the KG construction task within the knowledge domain, and ensure that the resulting KG aligns with the intended use cases. This also allows further ontology expansion by incorporating user-submitted domain- defining questions when interacting with the knowledge base, which serves as a user friendly interface of refining ontology by submitting new CQs and use our proposed pipeline to attach the incremental knowledge scope to existing ontology. 3.2. Relation Extraction and Ontology Matching During our preliminary experiments of prompting LLMs to directly generate ontology on documents, we noted that the LLM spontaneously recalled Wikidata knowledge in response, consistent with previous works [15]. In our preliminary experiments, this behaviour also transfers to small 7B/14B models. Following this direction, in the second step we extract relations from CQs and match them against Wikidata properties to better elicit model memories on Wikidata when constructing and using ontology. We first prompt LLMs to extract properties from CQ and write brief description on usage of extracted properties, including their domain and range, following editing guidelines of Wikidata. To match these properties against existing entries in Wikidata ontology, we pre- populate a candidate property list with all Wikidata properties after filtering out properties related to external database/knowledge base IDs. These extracted properties are then matched against the candidate list by a vector similarity search between description of properties. The representation for property is sentence embedding constructed from description of properties, and the top 1 closest candidate is retrieved for each extracted property. This matching result between each pair of extracted property and matched top 1 candidate is then vetted by LLM to see if they are really semantically similar as a final deduplication step. If a match is validated, the candidate property is added to the final property list; otherwise, the newly minted property is kept in the final list if we allow expansion from the candidate property list derived from Wikidata, and discarded when the final property list is required to be a subset of candidate property list. The first scenario is suitable for cases when no prior schema is known for the domain and some new properties outside of common ontology are expected, whereas the latter is for a known target list of possible properties. 3.3. Ontology Formatting In the third stage, we use LLM to generate an OWL ontology based on the matched and newly created properties. We copy the description, domain and range field from all properties under Wikidata semantics. For new properties, LLM is prompted to infer and summarize classes for the domain and range of the relations to output a complete OWL ontology, following the format of copied Wikidata properties. This step ensures that the resulting KG is grounded in a formal, machine-readable ontology that captures relationships between entities, and close to the semantics of Wikidata for interoperability. 3.4. KG Construction In the final stage, we use the LLM to construct a KG based on the CQs and related answers grounded by the generated ontology in the previous stage. For each (CQ, answer) pair, LLM extracts relevant entities and maps them to the ontology using the defined properties. The output is a set of RDF triples that constitutes the final KG. 4. Experiments and Discussion 4.1. Experiment settings We evaluate our ontology-grounded approach to KG construction (KGC) on three datasets for KGC datasets: Wiki-NRE [16], SciERC [17], and WebNLG [18]. As Wiki-NRE and WebNLG are partially based on Wikidata and DBpedia (derived from Wikipedia contents), and in our proposed pipeline, Wikidata schema is utilized, we include SciERC for a more robust evaluation, since SciERC contains relation types that are not equivalent by nature to properties in Wikidata. We used a subset Wiki-NRE’s test dataset containing 1,000 samples with 45 relation types following the split in [19], due to cost constraints. SciERC’s test set contain 974 samples under a schema with 7 relation types. For WebNLG, we used test set in Semantic Parsing (SP) task, with 1,165 samples and 159 relation types. For evaluation, we adopt partial F1 on KG triplets based on standards in [18]. All experiments are conducted for one-pass We note in the previous reports that annotation in KGC reports may be incomplete in terms of both possible relation types and KG triplets [19], [20]. As our pipeline is designed to autonomously uncover knowledge structure with no prior assumption on knowledge schema, we report our result in two ways, corresponding to the two configurations of final de-duplication step in Section 3.2: 1. Target schema constrained: In this setting, we match all relation types in test sets to its closest equivalent in Wikidata and constrict ontology to the relation universe in test set. 2. No schema constraint: In this setting, we do not filter matched ontology, even if they are not in schema of test dataset. This setting is close to real-life applications when processing documents with unknown schema. For property conjunction, evaluate for, compare, feature of in SciERC, we select the closest properties proposed by LLM based on our subjective opinion. To highlight our system’s competency, rather than directly prompting triplets, we parse output KG with RDF parser and extract all valid RDF triples for KG related to each document in test set, and present triplets to evaluation script for assessment. This ensures that our evaluation is on the generated KG ready to be consumed in downstream application. We test our pipeline on both Mistral-7B-instruct [21] and GPT-4o1 . Due to cost constraints, we have only tested GPT-4o on target schema constrained setting. For embedding property usage comment, we select bge-small-en [22]. We use GenIE [23], PL-Marker [24], and ReGen [25] as fine-tuned baseline for Wiki-NRE, SciERC, and WebNLG dataset, respectively (collectively named Non-LLM Baseline). For LLM-based systems, we use results reported in [19] for Wiki- NRE and WebNLG on the same Mistral model, and GPT-4 results in [3] for SciERC. (collectively named LLM Baseline). We note that it is highly unlikely that Mistral-7B poses an advantage over an earlier version of GPT-4, when interpreting result of SciERC. 4.2. Result 1 https://openai.com/index/hello-gpt-4o/ Table 1 shows the performance of our Method Wiki-NRE SciERC WebNLG method compared to state-of-the-art Non-LLM Baseline 0.484 0.532 0.767 baselines on this subset. Our proposed LLM Baseline 0.647 0.07 0.728 approach exceeds all baseline under tar- Proposed (Mistral) 0.66/0.60 0.73/0.58 0.74/0.68 get schema constrained setting on Wiki- Proposed (GPT-4o) 0.71/N/A 0.77/N/A 0.76/N/A NRE and SciERC datasets, while display- ing a small performance regression when Table 1: Partial F1 scores on test datasets. Best result without schema constraint. On WebNLG is bolded. Results of proposed pipeline under dataset, our pipeline maintained compet- two settings are presented as Target schema itiveness against fine-tuned SOTA when constrained/no schema constraint. constrained on target schema. These re- sults validate the quality of KG generated by our pipeline, especially SciERC, whose semantics contains properties that are not native to Wikidata. We also note performance improvement when using GPT-4o. 4.3. Discussion 4.3.1. Performance discrepancy on different grounding ontology It is worth noting that the relatively lower performance on no schema constraint setting across all datasets is due to the fact that the LLM discovers a richer ontology than the predefined target schema. While this expanded schema may capture additional relevant information, it can hinder extraction performance when evaluated solely against the limited target schema. This showcases the trade-off between schema completeness and strict adherence to a predefined ontology, and our pipeline performs best on a large set of documents with a limited scope of knowledge, requiring a concise schema. Furthermore, the flipside of performance deficit in an absence of schema constraints, i.e. additional ontology entries outside of dataset-defined properties, cannot be evaluated against the dataset directly, as the ontology is not entirely covered by test set annotations. Hence, the virtue of no schema constraint setting is to demonstrate that our pipeline can indeed provide a coverage of the properties in test set, though somewhat limited compared to baselines, when also capturing ontology outside test set schema, which is potentially more useful when discovering ontology on a novel document set with no expert knowledge in its schema conposition. This ability may be validated by manual evaluation on the full set of captured ontology in a future work down the line. Nevertheless, the marginal performance deficit leaves room for improvement. Recent reports explored that long input context may pose challenge to LLMs even if such long context length is technically supported [26]. We conjecture that aside from trimming grounding ontology, which hinders the knowledge coverage of our pipeline, few-shot fine-tuning on the new ontology or general pretraining in KG construction task may be helpful. We leave these as possible future directions. 4.3.2. Utility of generated KG It should be emphasized that, while the selected evaluation tasks evaluate the correctness of extracted triplets, the extracted knowledge graph can do more than that. With ongoing discussion related to grounding LLM knowledge on trusted knowledge sources to reduce hallucination [6], explicitly generating KG provides a path to audit knowledge elicited when interacting with LLM, and with evidence demonstrating that LLM has the potential to reason on graph and generate an explicit path to retrieve required knowledge [27], our pipeline may serve as a foundation for an interpretable QA system, where an LLM autonomously extracts ontology and deduces correct retrieval query based on the ontology when handling a set of unstructured document. The interpretability arise from the fact that KG and query could be understood and verified by users. Moreover, our usage of Wikidata schema offers potential interoperability with the whole Wikidata knowledge base, which safely expands the knowledge scope of QA system. We propose to continue research on this significant direction. 4.3.3. Computational resources We note the growing concern of sustainability in LLM applications due to intensive requirement on computational resources. This pipeline consumes three separate LLM calls per document, plus one call per extracted relation. It is not straight forward to compare the carbon footprint of our approach compared to Non-LLM baselines, as our work at this stage does not require model fine-tuning, whereas all of the Non-LLM baselines employed various tuning techniques when producing the result. On the other hand, our smallest model adopted, Mistral-7B, is more than 10x larger in terms of parameter size compared to T5 models used in Non-LLM baselines. Larger models naturally require more powerful GPU clusters in terms of both GPU quantity and capability, but our zero-shot approach may provide an advantage in terms of resource cost compared to Non-LLM baselines when processing a small number of documents with no training requirement. When comparing with LLM baselines, we note that the approach by [14], [3] consumes 1 and 2 LLM calls per document, respectively. However, we note that these baselines treat knowledge triplet as evaluation target, while we generate a formatted ontology at the end, which is more useful. Nevertheless, we recognize the performance burden and propose to explore techniques in fine-tuning and guided decoding to achieve better performance with smaller model and better reproducibility. 5. Conclusion We have demonstrated the effectiveness of our ontology-grounded approach to KG construction using LLMs. By leveraging the structured knowledge in Wikidata, pretrained on LLM, and grounding KG construction with generated ontology, our pipeline is able to construct high- quality KGs across various domains while maintaining competitive performance with state-of- the-art baselines. Generated KGs that are conformant with Wikidata schema leaves possibly wide open, of building an interpretable QA system that has robust access to both common knowledge and proprietary knowledge base. Acknowledgments This work is supported by Centre for Perceptual and Interactive Intelligence (CPII) Ltd, a CUHK-led InnoCentre under InnoHK scheme of Innovation and Technology Commision. References [1] A. Hogan, E. Blomqvist, M. Cochez, C. D’amato, G. D. Melo, C. Gutierrez, S. Kirrane, J. E. L. Gayo, R. Navigli, S. Neumaier, A.-C. N. Ngomo, A. Polleres, S. M. Rashid, A. Rula, L. Schmelzeisen, J. Sequeda, S. Staab, A. Zimmermann, Knowledge graphs, ACM Comput. Surv. 54 (2021). URL: https://doi.org/10.1145/3447772. doi:10.1145/3447772. [2] S. Ji, S. Pan, E. Cambria, P. Marttinen, P. S. Yu, A survey on knowledge graphs: Represen- tation, acquisition, and applications, IEEE Transactions on Neural Networks and Learning Systems 33 (2020) 494–514. URL: https://api.semanticscholar.org/CorpusID:211010433. [3] Y. Zhu, X. Wang, J. Chen, S. Qiao, Y. Ou, Y. Yao, S. Deng, H. Chen, N. Zhang, LLMs for Knowledge Graph Construction and Reasoning: Recent Capabilities and Future Opportu- nities, 2024. doi:10.48550/arXiv.2305.13168. arXiv:2305.13168. [4] F. Petroni, T. RocktΓ€schel, P. Lewis, A. Bakhtin, Y. Wu, A. H. Miller, S. Riedel, Language models as knowledge bases?, in: Conference on Empirical Methods in Natural Language Processing, 2019. URL: https://api.semanticscholar.org/CorpusID:202539551. [5] OpenAI, J. Achiam, S. Adler, S. Agarwal, L. Ahmad, I. Akkaya, F. L. Aleman, D. Almeida, J. Altenschmidt, S. Altman, S. Anadkat, R. Avila, I. Babuschkin, S. Balaji, V. Balcom, P. Baltescu, H. Bao, M. Bavarian, J. Belgum, I. Bello, J. Berdine, G. Bernadett-Shapiro, C. Berner, L. Bogdonoff, O. Boiko, M. Boyd, A.-L. Brakman, G. Brockman, T. Brooks, M. Brundage, K. Button, T. Cai, R. Campbell, A. Cann, B. Carey, C. Carlson, R. Carmichael, B. Chan, C. Chang, F. Chantzis, D. Chen, S. Chen, R. Chen, J. Chen, M. Chen, B. Chess, C. Cho, C. Chu, H. W. Chung, D. Cummings, J. Currier, Y. Dai, C. Decareaux, T. Degry, N. Deutsch, D. Deville, A. Dhar, D. Dohan, S. Dowling, S. Dunning, A. Ecoffet, A. Eleti, T. Eloundou, D. Farhi, L. Fedus, N. Felix, S. P. Fishman, J. Forte, I. Fulford, L. Gao, E. Georges, C. Gibson, V. Goel, T. Gogineni, G. Goh, R. Gontijo-Lopes, J. Gordon, M. Grafstein, S. Gray, R. Greene, J. Gross, S. S. Gu, Y. Guo, C. Hallacy, J. Han, J. Harris, Y. He, M. Heaton, J. Heidecke, C. Hesse, A. Hickey, W. Hickey, P. Hoeschele, B. Houghton, K. Hsu, S. Hu, X. Hu, J. Huizinga, S. Jain, S. Jain, J. Jang, A. Jiang, R. Jiang, H. Jin, D. Jin, S. Jomoto, B. Jonn, H. Jun, T. Kaftan, Łukasz Kaiser, A. Kamali, I. Kanitscheider, N. S. Keskar, T. Khan, L. Kilpatrick, J. W. Kim, C. Kim, Y. Kim, J. H. Kirchner, J. Kiros, M. Knight, D. Kokotajlo, Łukasz Kondraciuk, A. Kondrich, A. Konstantinidis, K. Kosic, G. Krueger, V. Kuo, M. Lampe, I. Lan, T. Lee, J. Leike, J. Leung, D. Levy, C. M. Li, R. Lim, M. Lin, S. Lin, M. Litwin, T. Lopez, R. Lowe, P. Lue, A. Makanju, K. Malfacini, S. Manning, T. Markov, Y. Markovski, B. Martin, K. Mayer, A. Mayne, B. McGrew, S. M. McKinney, C. McLeavey, P. McMillan, J. McNeil, D. Medina, A. Mehta, J. Menick, L. Metz, A. Mishchenko, P. Mishkin, V. Monaco, E. Morikawa, D. Mossing, T. Mu, M. Murati, O. Murk, D. MΓ©ly, A. Nair, R. Nakano, R. Nayak, A. Neelakantan, R. Ngo, H. Noh, L. Ouyang, C. O’Keefe, J. Pachocki, A. Paino, J. Palermo, A. Pantuliano, G. Parascandolo, J. Parish, E. Parparita, A. Passos, M. Pavlov, A. Peng, A. Perelman, F. de Avila Belbute Peres, M. Petrov, H. P. de Oliveira Pinto, Michael, Pokorny, M. Pokrass, V. H. Pong, T. Powell, A. Power, B. Power, E. Proehl, R. Puri, A. Radford, J. Rae, A. Ramesh, C. Raymond, F. Real, K. Rimbach, C. Ross, B. Rotsted, H. Roussez, N. Ryder, M. Saltarelli, T. Sanders, S. Santurkar, G. Sastry, H. Schmidt, D. Schnurr, J. Schulman, D. Selsam, K. Sheppard, T. Sherbakov, J. Shieh, S. Shoker, P. Shyam, S. Sidor, E. Sigler, M. Simens, J. Sitkin, K. Slama, I. Sohl, B. Sokolowsky, Y. Song, N. Staudacher, F. P. Such, N. Summers, I. Sutskever, J. Tang, N. Tezak, M. B. Thompson, P. Tillet, A. Tootoonchian, E. Tseng, P. Tuggle, N. Turley, J. Tworek, J. F. C. Uribe, A. Vallone, A. Vijayvergiya, C. Voss, C. Wainwright, J. J. Wang, A. Wang, B. Wang, J. Ward, J. Wei, C. Weinmann, A. Welihinda, P. Welinder, J. Weng, L. Weng, M. Wiethoff, D. Willner, C. Winter, S. Wolrich, H. Wong, L. Workman, S. Wu, J. Wu, M. Wu, K. Xiao, T. Xu, S. Yoo, K. Yu, Q. Yuan, W. Zaremba, R. Zellers, C. Zhang, M. Zhang, S. Zhao, T. Zheng, J. Zhuang, W. Zhuk, B. Zoph, Gpt-4 technical report, 2024. arXiv:2303.08774. [6] G. Agrawal, T. Kumarage, Z. Alghamdi, H. Liu, Can Knowledge Graphs Reduce Hallucinations in LLMs? : A Survey, 2024. doi:10.48550/arXiv.2311.07914. arXiv:2311.07914. [7] E. Agichtein, L. Gravano, Snowball: extracting relations from large plain-text collections, in: Proceedings of the Fifth ACM Conference on Digital Libraries, DL ’00, Association for Computing Machinery, New York, NY, USA, 2000, p. 85–94. URL: https://doi.org/10.1145/ 336597.336644. doi:10.1145/336597.336644. [8] Y. Zhang, V. Zhong, D. Chen, G. Angeli, C. D. Manning, Position-aware attention and supervised data improve slot filling, in: M. Palmer, R. Hwa, S. Riedel (Eds.), Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics, Copenhagen, Denmark, 2017, pp. 35–45. URL: https:// aclanthology.org/D17-1004. doi:10.18653/v1/D17-1004. [9] M. Mintz, S. Bills, R. Snow, D. Jurafsky, Distant supervision for relation extraction without labeled data, in: K.-Y. Su, J. Su, J. Wiebe, H. Li (Eds.), Proceedings of the Joint Conference of the 47th Annual Meeting of the ACL and the 4th International Joint Conference on Natural Language Processing of the AFNLP, Association for Computational Linguistics, Suntec, Singapore, 2009, pp. 1003–1011. URL: https://aclanthology.org/P09-1113. [10] X. L. Dong, E. Gabrilovich, G. Heitz, W. Horn, N. Lao, K. Murphy, T. Strohmann, S. Sun, W. Zhang, Knowledge vault: A web-scale approach to probabilistic knowledge fusion, in: The 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD ’14, New York, NY, USA - August 24 - 27, 2014, 2014, pp. 601–610. URL: http: //www.cs.cmu.edu/~nlao/publication/2014.kdd.pdf, evgeniy GabrilovichWilko HornNi LaoKevin MurphyThomas StrohmannShaohua SunWei ZhangGeremy Heitz. [11] A. Bosselut, H. Rashkin, M. Sap, C. Malaviya, A. Celikyilmaz, Y. Choi, COMET: Com- monsense transformers for automatic knowledge graph construction, in: A. Korhonen, D. Traum, L. MΓ rquez (Eds.), Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics, Florence, Italy, 2019, pp. 4762–4779. URL: https://aclanthology.org/P19-1470. doi:10.18653/v1/P19-1470. [12] Y. Chen, Y. Liu, L. Dong, S. Wang, C. Zhu, M. Zeng, Y. Zhang, AdaPrompt: Adap- tive model training for prompt-based NLP, in: Y. Goldberg, Z. Kozareva, Y. Zhang (Eds.), Findings of the Association for Computational Linguistics: EMNLP 2022, Associ- ation for Computational Linguistics, Abu Dhabi, United Arab Emirates, 2022, pp. 6057– 6068. URL: https://aclanthology.org/2022.findings-emnlp.448. doi:10.18653/v1/2022. findings-emnlp.448. [13] S. Yu, T. He, J. Glass, AutoKG: Constructing Virtual Knowledge Graphs from Unstruc- tured Documents for Question Answering, 2021. doi:10.48550/arXiv.2008.08995. arXiv:2008.08995. [14] B. Chen, A. L. Bertozzi, AutoKG: Efficient Automated Knowledge Graph Generation for Language Models, 2023. doi:10.48550/arXiv.2311.14740. arXiv:2311.14740. [15] S. J. Semnani, V. Z. Yao, H. C. Zhang, M. S. Lam, WikiChat: Stopping the Hallucina- tion of Large Language Model Chatbots by Few-Shot Grounding on Wikipedia, 2023. arXiv:2305.14292. [16] B. D. Trisedya, G. Weikum, J. Qi, R. Zhang, Neural relation extraction for knowledge base enrichment, in: A. Korhonen, D. Traum, L. MΓ rquez (Eds.), Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Association for Computational Linguistics, Florence, Italy, 2019, pp. 229–240. URL: https://aclanthology.org/P19-1023. doi:10.18653/v1/P19-1023. [17] Y. Luan, L. He, M. Ostendorf, H. Hajishirzi, Multi-task identification of entities, relations, and coreferencefor scientific knowledge graph construction, in: Proc. Conf. Empirical Methods Natural Language Process. (EMNLP), 2018. [18] T. Castro Ferreira, C. Gardent, N. Ilinykh, C. van der Lee, S. Mille, D. Moussallem, A. Shi- morina, The 2020 bilingual, bi-directional WebNLG+ shared task: Overview and evalua- tion results (WebNLG+ 2020), in: T. Castro Ferreira, C. Gardent, N. Ilinykh, C. van der Lee, S. Mille, D. Moussallem, A. Shimorina (Eds.), Proceedings of the 3rd International Workshop on Natural Language Generation from the Semantic Web (WebNLG+), Asso- ciation for Computational Linguistics, Dublin, Ireland (Virtual), 2020, pp. 55–76. URL: https://aclanthology.org/2020.webnlg-1.7. [19] B. Zhang, H. Soh, Extract, Define, Canonicalize: An LLM-based Framework for Knowledge Graph Construction, 2024. arXiv:2404.03868. [20] R. Han, T. Peng, C. Yang, B. Wang, L. Liu, X. Wan, Is information extraction solved by chatgpt? an analysis of performance, evaluation criteria, robustness and errors, 2023. arXiv:2305.14450. [21] A. Q. Jiang, A. Sablayrolles, A. Mensch, C. Bamford, D. S. Chaplot, D. de las Casas, F. Bres- sand, G. Lengyel, G. Lample, L. Saulnier, L. R. Lavaud, M.-A. Lachaux, P. Stock, T. L. Scao, T. Lavril, T. Wang, T. Lacroix, W. E. Sayed, Mistral 7b, 2023. arXiv:2310.06825. [22] J. Chen, S. Xiao, P. Zhang, K. Luo, D. Lian, Z. Liu, Bge m3-embedding: Multi-lingual, multi-functionality, multi-granularity text embeddings through self-knowledge distillation, 2023. arXiv:2309.07597. [23] M. Josifoski, N. De Cao, M. Peyrard, F. Petroni, R. West, GenIE: Generative information extraction, in: M. Carpuat, M.-C. de Marneffe, I. V. Meza Ruiz (Eds.), Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Association for Computational Linguistics, Seattle, United States, 2022, pp. 4626–4643. URL: https://aclanthology.org/2022.naacl-main. 342. doi:10.18653/v1/2022.naacl-main.342. [24] D. Ye, Y. Lin, P. Li, M. Sun, Packed levitated marker for entity and relation extraction, in: S. Muresan, P. Nakov, A. Villavicencio (Eds.), Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), Association for Computational Linguistics, Dublin, Ireland, 2022, pp. 4904–4917. URL: https://aclanthology. org/2022.acl-long.337. doi:10.18653/v1/2022.acl-long.337. [25] P. Dognin, I. Padhi, I. Melnyk, P. Das, ReGen: Reinforcement learning for text and knowledge base generation using pretrained language models, in: M.-F. Moens, X. Huang, L. Specia, S. W.-t. Yih (Eds.), Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, Association for Computational Linguistics, Online and Punta Cana, Dominican Republic, 2021, pp. 1084–1099. URL: https://aclanthology.org/2021. emnlp-main.83. doi:10.18653/v1/2021.emnlp-main.83. [26] T. Li, G. Zhang, Q. D. Do, X. Yue, W. Chen, Long-context llms struggle with long in-context learning, 2024. arXiv:2404.02060. [27] F. Brei, J. Frey, L.-P. Meyer, Leveraging small language models for text2sparql tasks to improve the resilience of ai assistance, 2024. arXiv:2405.17076. A. Sample generated KG This KG was generated under no schema constraint setting for this document: Mohammad Firouzi ( Born 1958 Tehran ) is a prolific Iranian musician , whose primary instrument is the barbat . < P r e f i x e s and d e f i n i t i o n o f d e p e n d e n c i e s o m i t t e d > wd : Mohammad_Firouzi a wd : human ; r d f s : l a b e l " Mohammad F i r o u z i " @en ; wdt : o c c u p a t i o n wd : M u s i c i a n ; wdt : C o u n t r y O f C i t i z e n s h i p wd : I r a n ; wdt : P l a c e O f B i r t h wd : Tehran ; wdt : D a t e O f B i r t h " 1 9 5 8 " ^ ^ x s d : d a t e . Note that in official annotation, only triplets related to place of birth and nationality exist, hence the evaluation will be penalized with low precision. B. Preprocessing of Wikidata schema To save space in LLM input context and mitigate performance drop on selected target schema when ontology is large, we only include commonly used properties by restricting data type on item, quantity, string, monolingual text, point in time.2 To align with common pretraining objectives of LLM, we substitute entity identifiers (e.g. P19) with its literal label (rdfs:label in PascalCase (e.g. PlaceOfBirth). 2 https://www.wikidata.org/wiki/Help:Data_type C. Prompts All prompts are reused across all datasets. C.1. CQ generation We prompt LLM to generate up to 3 CQs per document for efficiency considering nature of test datasets, but note that this may be adjusted. W r i t e competency q u e s t i o n s b a s e d on t h e a b s t r a c t l e v e l c o n c e p t s i n t h e document . W r i t e q u e s t i o n s t h a t can be a nswered u s i n g t h e document o n l y . W r i t e up t o 3 q u e s t i o n s p e r document . Below a r e t h e e x a m p l e s and f o l l o w t h e same f o r m a t when g e n e r a t i n g competency q u e s t i o n s : #### Document : D o u g l a s Noel Adams ( 1 1 March 1 9 5 2 βˆ’ 11 May 2 0 0 1 ) was an E n g l i s h a u t h o r , h u m o u r i s t , and s c r e e n w r i t e r , b e s t known f o r The H i t c h h i k e r ’ s Guide t o t h e G a l a x y (HHGTTG) . O r i g i n a l l y a 1 9 7 8 BBC r a d i o comedy , The H i t c h h i k e r ’ s Guide t o t h e G a l a x y d e v e l o p e d i n t o a " t r i l o g y " o f f i v e b oo ks t h a t s o l d more t h a n 15 m i l l i o n c o p i e s i n h i s l i f e t i m e . I t was f u r t h e r developed into a t e l e v i s i o n series , s e v e r a l stage p l a y s , comics , a v i d e o game , and a 2 0 0 5 f e a t u r e f i l m . Adams ’ s c o n t r i b u t i o n t o UK r a d i o i s commemorated i n The R a d i o Academy ’ s H a l l o f Fame . #### Questions : CQ1 . What i s t h e d a t e o f b i r t h o f D o u g l a s Noel Adams ? CQ2 . What i s t h e d a t e o f d e a t h o f D o u g l a s Noel Adams ? CQ3 . What i s t h e o c c u p a t i o n o f D o u g l a s Noel Adams ? CQ4 . What i s t h e c o u n t r y o f c i t i z e n s h i p o f D o u g l a s Noel Adams ? CQ5 . What i s t h e most n o t a b l e work o f D o u g l a s Noel Adams ? CQ6 . What i s t h e o r i g i n a l medium o f The H i t c h h i k e r ’ s Guide t o the Galaxy ? CQ7 . I n what y e a r was The H i t c h h i k e r ’ s Guide t o t h e G a l a x y originally broadcast ? CQ8 . How many bo ok s a r e i n The H i t c h h i k e r ’ s Guide t o t h e G a l a x y " trilogy "? CQ9 . What o t h e r media a d a p t a t i o n s were c r e a t e d b a s e d on The H i t c h h i k e r ’ s Guide t o t h e G a l a x y ? #### Document : { document t o be p r o c e s s e d } #### Questions : C.2. CQ answering Use t h e p r o v i d e d document t o answer u s e r q u e r y . I f you don ’ t know t h e answer , j u s t s a y t h a t you don ’ t know , don ’ t t r y t o make up an answer . P a s s a g e : { doc } Query : { q u e r y } C.3. Relation extraction You a r e an a s s i s t a n t i n b u i l d i n g a knowledge g r a p h . A n a l y z e t h e f o l l o w i n g competency q u e s t i o n s and i d e n t i f y a l l r e l a t i o n s h i p s and c o n c e p t s c o n c e p t s m e n t i o n e d i n t h e question . E x t r a c t r e l a t i o n f i r s t , then d e s c r i b e the usage of each r e l a t i o n b a s e d on your u n d e r s t a n d i n g g i v e n t h e c o n t e x t o f competency q u e s t i o n s . Afterwards , e x t r a c t a l l r e l a t i o n βˆ’ r e l a t e d concepts . You s h o u l d o n l y e x t r a c t p r o p e r t i e s between e n t i t i e s and l i t e r a l s , not e n t i t i e s themselves , or c l a s s e s of e n t i t i e s . T h e r e f o r e , n o t a l l CQs c o n t a i n v a l i d p r o p e r t i e s . I f you don ’ t know t h e answer , j u s t s a y t h a t you don ’ t know , don ’ t t r y t o make up an answer . Merge a l l r e l a t i o n s i n t o one l i s t and a l l c o n c e p t s i n t o one list . Do n o t r e p l y u s i n g a c o m p l e t e s e n t e n c e , and o n l y g i v e t h e answer i n t h e f o l l o w i n g f o r m a t . Below a r e t h e e x a m p l e s and f o l l o w t h e same f o r m a t t o e x t r a c t the r e l a t i o n s : #### Document : D o u g l a s Noel Adams ( 1 1 March 1 9 5 2 βˆ’ 11 May 2 0 0 1 ) was an E n g l i s h a u t h o r , h u m o u r i s t , and s c r e e n w r i t e r , b e s t known f o r The H i t c h h i k e r ’ s Guide t o t h e G a l a x y (HHGTTG) . O r i g i n a l l y a 1 9 7 8 BBC r a d i o comedy , The H i t c h h i k e r ’ s Guide t o t h e G a l a x y d e v e l o p e d i n t o a " t r i l o g y " o f f i v e b oo ks t h a t s o l d more t h a n 15 m i l l i o n c o p i e s i n h i s l i f e t i m e . I t was f u r t h e r developed into a t e l e v i s i o n series , s e v e r a l stage p l a y s , comics , a v i d e o game , and a 2 0 0 5 f e a t u r e f i l m . Adams ’ s c o n t r i b u t i o n t o UK r a d i o i s commemorated i n The R a d i o Academy ’ s H a l l o f Fame . #### Questions : CQ1 . What i s t h e d a t e o f b i r t h o f D o u g l a s Noel Adams ? CQ2 . What i s t h e d a t e o f d e a t h o f D o u g l a s Noel Adams ? CQ3 . What i s t h e o c c u p a t i o n o f D o u g l a s Noel Adams ? CQ4 . What i s t h e c o u n t r y o f c i t i z e n s h i p o f D o u g l a s Noel Adams ? CQ5 . What i s t h e most n o t a b l e work o f D o u g l a s Noel Adams ? CQ6 . What i s t h e o r i g i n a l medium o f The H i t c h h i k e r ’ s Guide t o the Galaxy ? CQ7 . I n what y e a r was The H i t c h h i k e r ’ s Guide t o t h e G a l a x y originally broadcast ? CQ8 . How many bo ok s a r e i n The H i t c h h i k e r ’ s Guide t o t h e G a l a x y " trilogy "? CQ9 . What o t h e r media a d a p t a t i o n s were c r e a t e d b a s e d on The H i t c h h i k e r ’ s Guide t o t h e G a l a x y ? #### Relations : ( d a t e o f b i r t h , The d a t e on which t h e s u b j e c t was born . ) ( d a t e o f d e a t h , The d a t e on which t h e s u b j e c t d i e d . ) ( o c c u p a t i o n , The o c c u p a t i o n o f a p e r s o n . ) ( c o u n t r y o f c i t i z e n s h i p , The c o u n t r y o f which t h e s u b j e c t i s a citizen .) ( n o t a b l e work , The most n o t a b l e work o f a p e r s o n . ) ( genre , The g e n r e o r t y p e o f work . ) ( p u b l i c a t i o n d a t e , The d a t e o r p e r i o d when a work was f i r s t p u b l i s h e d or r e l e a s e d . ) ( has part , I n d i c a t e s t h a t the s u b j e c t has a c e r t a i n part , component , o r e l e m e n t . ) ( s e r i e s , I n d i c a t e s t h a t the s u b j e c t i s p a r t of a s e r i e s , such a s a book s e r i e s , f i l m s e r i e s , o r t e l e v i s i o n s e r i e s . ) #### Document : { document t o be p r o c e s s e d } #### Questions : { CQs } #### Relations : C.4. Ontology matching D e c i d e i f t h e two p r o p e r t i e s a r e s e m a n t i c a l l y s i m i l a r i n an ontology . You s h o u l d s a y y e s i f you d e c i d e t h a t t h e s e p r o p t i e s a r e s i m i l a r , or i f they are i n v e r s e p r o p e r t i e s . Answer i n " y e s " o r " no " o n l y . P r o p e r t y 1 : { p1 } P r o p e r t y 2 : { p2 } C.5. Ontology formatting For properties under Wikidata schema, we retrieve schema:description, rdfs:domain, rdfs:range for each property and include it in resulting ontology. Otherwise LLM is prompted to author ontology as so: Use t h e r e l a t i o n s ( p r o p e r t i e s ) and t h e i r u s a g e comments t o b u i l d an o n t o l o g y i n RDF f o r m a t . I f you don ’ t know t h e answer , j u s t s a y t h a t you don ’ t know , don ’ t t r y t o make up an answer . Don ’ t p r o v i d e a n y t h i n g o t h e r t h a n an o n t o l o g y i n RDF f o r m a t . I n f e r and summarize c l a s s e s f o r domain and r a n g e o f t h e r e l a t i o n s a c r o s s t h e c o n c e p t s p r o v i d e d , and add t h e s e c l a s s e s to r e l a t i o n s only i f r e q u i r e d f o r c l o u s r e of relations . F o r e a c h r e l a t i o n , add r e l e v a n t o n t o l o g y e n t r y f o r i t . Add r d f s : comment b a s e d on t h e u s a g e comments . Use wdt : namespace f o r a l l r e l a t i o n s d i s c o v e r e d . Use e n t i t i e s un d e r t h e s e p r e f i x e s i f n e c e s s a r y : @ p r e f i x r d f : < h t t p : / / www. w3 . o r g / 1 9 9 9 / 0 2 / 2 2 βˆ’ r d f βˆ’ s y n t a x βˆ’ ns #> . @ p r e f i x x s d : < h t t p : / / www. w3 . o r g / 2 0 0 1 / XMLSchema#> . @ p r e f i x r d f s : < h t t p : / / www. w3 . o r g / 2 0 0 0 / 0 1 / r d f βˆ’ schema #> . @ p r e f i x owl : < h t t p : / / www. w3 . o r g / 2 0 0 2 / 0 7 / owl #> . @ p r e f i x w i k i b a s e : < h t t p : / / w i k i b a . s e / o n t o l o g y #> . @ p r e f i x schema : < h t t p : / / schema . o r g / > . @ p r e f i x wd : < h t t p : / / www. w i k i d a t a . o r g / e n t i t y / > . @ p r e f i x wdt : < h t t p : / / www. w i k i d a t a . o r g / prop / d i r e c t / > . Use t u r t l e s y n t a x . Below i s an example : #### Relations : ( r e s u l t s , r e s u l t s : r e s u l t s of a c o m p e t i t i o n such as s p o r t s or elections ) #### Ontology : wdt : R e s u l t s a w i k i b a s e : P r o p e r t y ; schema : d e s c r i p t i o n " r e s u l t s o f a c o m p e t i t i o n s u c h a s s p o r t s or e l e c t i o n s " ; rdfs : label " results " ; r d f s : domain wd : r e f e r e n d u m , wd : c o m p e t i t i o n , wd : p a r t y c o n f e r e n c e , wd : s p o r t i n g e v e n t ; r d f s : r a n g e wd : e l e c t o r a l r e s u l t , wd : v o t i n g r e s u l t , wd : s p o r t r e s u l t , wd : r a c e r e s u l t . #### Relations : { relation } #### Ontology : C.6. KG generation Your t a s k i s t o c o n s t r u c t a knowledge g r a p h b a s e d on t h e provided ontology . F o c u s on u n d e r s t a n d i n g r e l a t i o n s h i p s from t h e q u e s t i o n answer p a i r and document , and e x t r a c t r e l a t e d e n t i t i e s , t h e n mapping them t o t h e o n t o l o g y using the p r o p e r t i e s defined in the ontology . Do n o t i n c l u d e new p r o p e r t i e s o t h e r t h a n t h o s e i n o n t o l o g y . Only u s e t h o s e p r o p e r t i e s i n t h e o n t o l o g y . Output i n t u r t l e f o r m a t f o l l o w i n g t h e o n t o l o g y p r o v i d e d . You s h o u l d o n l y i n c l u d e knowledge i n q u e s t i o n answer p a i r s and t h e document . Do n o t make up a n s w e r s . Use t h i s o n t o l o g y b a s e d on W i k i d a t a a s t h e s t a r t i n g p o i n t : { ont } Below i s an example : #### Document : D o u g l a s Noel Adams ( 1 1 March 1 9 5 2 βˆ’ 11 May 2 0 0 1 ) was an E n g l i s h a u t h o r , h u m o u r i s t , and s c r e e n w r i t e r , b e s t known f o r The H i t c h h i k e r ’ s Guide t o t h e G a l a x y (HHGTTG) . O r i g i n a l l y a 1 9 7 8 BBC r a d i o comedy , The H i t c h h i k e r ’ s Guide t o t h e G a l a x y d e v e l o p e d i n t o a " t r i l o g y " o f f i v e b oo ks t h a t s o l d more t h a n 15 m i l l i o n c o p i e s i n h i s l i f e t i m e . I t was f u r t h e r d e v e l o p e d i n t o a t e l e v i s i o n s e r i e s , s e v e r a l s t a g e p l a y s , comics , a v i d e o game , and a 2 0 0 5 f e a t u r e f i l m . Adams ’ s c o n t r i b u t i o n t o UK r a d i o i s commemorated i n The R a d i o Academy ’ s H a l l o f Fame . #### Q u e s t i o n answer p a i r s : Q : What i s D o u g l a s Adams an i n s t a n c e o f ? A : D o u g l a s Adams i s an i n s t a n c e o f human . Q : What i s D o u g l a s Adams ’ s e x o r g e n d e r ? A : D o u g l a s Adams ’ s e x o r g e n d e r i s male . Q : Where was D o u g l a s Adams born ? A : D o u g l a s Adams was born i n Cambridge . Q : Where d i d D o u g l a s Adams d i e ? A : D o u g l a s Adams d i e d i n S a n t a B a r b a r a , C a l i f o r n i a . Q : When was D o u g l a s Adams born ? A : D o u g l a s Adams was born on 1 9 5 2 βˆ’ 0 3 βˆ’ 1 1 . Q : On what d a t e d i d D o u g l a s Adams d i e ? A : D o u g l a s Adams d i e d on 2 0 0 1 βˆ’ 0 5 βˆ’ 1 1 . Q : What o c c u p a t i o n d i d D o u g l a s Adams have ? A : D o u g l a s Adams was a w r i t e r , comedian , and d r a m a t i s t . Q : What l a n g u a g e s d i d D o u g l a s Adams speak , w r i t e , o r s i g n ? A : D o u g l a s Adams spoke , wrote , o r s i g n e d E n g l i s h . Q : Where was D o u g l a s Adams e d u c a t e d ? A : D o u g l a s Adams was e d u c a t e d a t S t John ’ s C o l l e g e , Cambridge and Brentwood S c h o o l , E s s e x . Q : What i n s t i t u t i o n i s D o u g l a s Adams an a l u m n i o f ? A : D o u g l a s Adams i s an a l u m n i o f S t John ’ s C o l l e g e . Q : What a r e some n o t a b l e works by D o u g l a s Adams ? A : Some n o t a b l e works by D o u g l a s Adams i n c l u d e The H i t c h h i k e r ’ s Guide t o t h e G a l a x y and D i r k G e n t l y ’ s H o l i s t i c D e t e c t i v e Agency . Q : Was D o u g l a s Adams a member o f any n o t a b l e o r g a n i z a t i o n s ? A : Yes , D o u g l a s Adams was a member o f Monty Python and The I n d e p e n d e n t on Sunday . Q : What award d i d D o u g l a s Adams r e c e i v e ? A : D o u g l a s Adams r e c e i v e d t h e L o c u s Award f o r B e s t S c i e n c e F i c t i o n Novel . Q : What i s t h e Commons C a t e g o r y f o r D o u g l a s Adams ? A : The Commons C a t e g o r y f o r D o u g l a s Adams i s " D o u g l a s Adams " . #### Ontology : @ p r e f i x r d f : h t t p : / / www. w3 . o r g / 1 9 9 9 / 0 2 / 2 2 βˆ’ r d f βˆ’ s y n t a x βˆ’ ns # . @ p r e f i x r d f s : h t t p : / / www. w3 . o r g / 2 0 0 0 / 0 1 / r d f βˆ’ schema # . @ p r e f i x wdt : h t t p : / / www. w i k i d a t a . o r g / prop / d i r e c t / . @ p r e f i x wd : h t t p : / / www. w i k i d a t a . o r g / e n t i t y / . @ p r e f i x x s d : h t t p : / / www. w3 . o r g / 2 0 0 1 / XMLSchema# . wd : Douglas_Adams r d f s : l a b e l " D o u g l a s Adams " @en ; wdt : I n s t a n c e O f wd : human ; wdt : SexOrGender wd : male ; wdt : P l a c e O f B i r t h wd : Cambridge ; wdt : P l a c e O f D e a t h wd : S a n t a _ B a r b a r a _ C a l i f o r n i a ; wdt : D a t e O f B i r t h " 1 9 5 2 βˆ’ 0 3 βˆ’ 1 1 " ^ ^ x s d : d a t e ; wdt : D a t e O f D e a t h " 2 0 0 1 βˆ’ 0 5 βˆ’ 1 1 " ^ ^ x s d : d a t e ; wdt : O c c u p a t i o n wd : w r i t e r ; wdt : O c c u p a t i o n wd : comedian ; wdt : O c c u p a t i o n wd : d r a m a t i s t ; wdt : L a n g u a g e s S p o k e n W r i t t e n O r S i g n e d wd : E n g l i s h ; wdt : E d u c a t e d A t wd : S t _ J o h n s _ C o l l e g e _ C a m b r i d g e ; wdt : E d u c a t e d A t wd : B r e n t w o o d _ S c h o o l _ E s s e x ; wdt : AlumniOf wd : S t _ J o h n s _ C o l l e g e ; wdt : NotableWork wd : T h e _ H i t c h h i k e r s _ G u i d e _ t o _ t h e _ G a l a x y ; wdt : NotableWork wd : D i r k _ G e n t l y s _ H o l i s t i c _ D e t e c t i v e _ A g e n c y ; wdt : MemberOf wd : Monty_Python ; wdt : M e m b e r O f O r g a n i z a t i o n wd : T h e _ I n d e p e n d e n t _ o n _ S u n d a y ; wdt : Award wd : L o c u s _ A w a r d _ f o r _ B e s t _ S c i e n c e _ F i c t i o n _ N o v e l ; wdt : CommonsCategory " D o u g l a s Adams " @en . #### Document : { doc } #### Q u e s t i o n s and Answer p a i r s : { qa } #### Ontology :