<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Towards a Gateway for Knowledge Graph Schemas Collection, Analysis, and Embedding</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Mattia</forename><surname>Fumagalli</surname></persName>
							<email>mattia.fumagalli@unibz.it</email>
							<affiliation key="aff0">
								<orgName type="institution">Free University of Bozen-Bolzano</orgName>
								<address>
									<settlement>Bolzano</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Marco</forename><surname>Boffo</surname></persName>
							<email>marco.boffo@studenti.unitn.it</email>
							<affiliation key="aff1">
								<orgName type="institution">Qascom SRL</orgName>
								<address>
									<settlement>Vicenza</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Daqian</forename><surname>Shi</surname></persName>
							<email>daqian.shi@unitn.it</email>
							<affiliation key="aff2">
								<orgName type="institution">University of Trento</orgName>
								<address>
									<settlement>Trento</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Mayukh</forename><surname>Bagchi</surname></persName>
							<email>mayukh.bagchi@unitn.it</email>
							<affiliation key="aff2">
								<orgName type="institution">University of Trento</orgName>
								<address>
									<settlement>Trento</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Fausto</forename><surname>Giunchiglia</surname></persName>
							<email>fausto.giunchiglia@unitn.it</email>
							<affiliation key="aff2">
								<orgName type="institution">University of Trento</orgName>
								<address>
									<settlement>Trento</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff3">
								<address>
									<postCode>2023</postCode>
									<settlement>Sherbrooke</settlement>
									<region>Québec</region>
									<country key="CA">Canada</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Towards a Gateway for Knowledge Graph Schemas Collection, Analysis, and Embedding</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">D5232A8AF53FF57033437338D1B643AC</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T19:26+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Knowledge graphs</term>
					<term>knowledge graphs catalog</term>
					<term>knowledge graphs analysis</term>
					<term>knowledge graphs embedding</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>One of the significant barriers to the training of statistical models on knowledge graphs is the difficulty that scientists have in finding the best input data to address their prediction goal. In addition to this, a key challenge is to determine how to manipulate these relational data, which are often in the form of particular triples (i.e., subject, predicate, object), to enable the learning process. Currently, many high-quality catalogs of knowledge graphs, are available. However, their primary goal is the re-usability of these resources, and their interconnection, in the context of the Semantic Web. This paper describes the LiveSchema initiative, namely, the first version of a gateway that has the main scope of leveraging the gold mine of data collected by many existing catalogs collecting relational data like ontologies and knowledge graphs. At the current state, LiveSchema contains ∼ 1000 datasets from 4 main sources and offers some key facilities, which allow to: i) evolving LiveSchema, by aggregating other source catalogs and repositories as input sources; ii) querying all the collected resources; iii) transforming each given dataset into formal concept analysis matrices that enable analysis and visualization services; iv) generating models and tensors from each given dataset.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Finding the best data to train statistical models and properly address the target learning goal is widely recognized as one of the most pivotal tasks in Machine Learning (ML) <ref type="bibr" target="#b0">[1]</ref>. ML models highly depend, indeed, on the quality of data they receive as input. While, so far, the development of highly efficient and scalable learning methods, to address critical prediction tasks (see, for instance, image classification and information patterns recognition <ref type="bibr" target="#b1">[2]</ref>), helped data scientists and analytics professionals in scaling their activities, the process of finding, selecting and improving the quality of these data still requires a considerable amount of time and manual effort <ref type="bibr" target="#b2">[3]</ref>. This latter challenge is also present when statistical models are trained on knowledge representations, like knowledge graphs and ontologies <ref type="bibr" target="#b3">[4]</ref>, where the data received as input are graph-structured data, consisting of entities (or nodes) and labeled links, or edges, (relations between entities). In this setting, the final learning goal can be identified as the prediction of missing relations between nodes, the prediction of nodes properties, and the clustering of the nodes based on their connections, these being common goals arising in many scenarios, such as analysis of social networks and biological pathways <ref type="bibr" target="#b4">[5]</ref>. Consequently, the efficacy of the ML algorithms directly depends on the quality of the input graphs, as well as their relevance to the domain of application.</p><p>In this paper, leveraging the ideas presented in <ref type="bibr" target="#b5">[6,</ref><ref type="bibr" target="#b6">7,</ref><ref type="bibr" target="#b7">8]</ref> and <ref type="bibr" target="#b8">[9]</ref>, where an approach to analyzing knowledge graph schemas to address Entity Type Recognition (ETR) tasks has been devised, we introduce the LiveSchema initiative, namely the first version of a gateway that has the main scope of exploiting the gold mine of data collected by many existing catalogs collecting relational data like ontologies and knowledge graphs. At the current state, LiveSchema contains ∼ 1000 datasets from 4 sources and offers some key facilities, which allow to: i) continuously updating a catalog of knowledge graphs, by aggregating other source catalogs and repositories; ii) querying all the collected resources; iii) transforming each given dataset into formal concept analysis matrices that enable analysis and visualization services; iv) generating models and tensors from each given dataset. At the current state, LiveSchema is accessible at http://liveschema.eu/ and it is ready to be demonstrated. The admin functionalities can be accessed and tested at http://liveschema.eu/user/login, by using 'reviewer' as admin/password. This demonstration paper is organized as follows. Section 2 introduces the motivation guiding the initiative. Section 3 gives a brief overview of the data architecture. Section 4 shows how LiveSchema can be evolved with new knowledge data. Section 4 illustrates the main LiveSchema components. Section 5 provides an example of how Liveschema can be used. Section 6 discusses some implications and limitations. Section 7 provides the conclusion.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Motivation</head><p>As an example scenario, suppose that a data scientist needs to run a standard Entity Type Recognition task (ETR), <ref type="foot" target="#foot_0">1</ref> as it is described in <ref type="bibr" target="#b7">[8]</ref> and <ref type="bibr" target="#b11">[12]</ref>, where the goal is to recognize objects of the type 'Person' across a set of multiple tabular data, coming, for instance, from an open data repository. This may involve that she needs to find a reference ontology containing: i. the target class and corresponding label; ii. possibly a huge number of properties for the target class, to increase the probability to match some of the input test properties; iii. possibly a low number of overlapping properties, in order to decrease the number of false-negative/positive predictions.</p><p>The process of searching, analyzing, and transforming the target ontology can take a long time and it may involve a considerable effort. The scientist has, indeed, to go through a broad search over the available resources and related catalogs, possibly checking multiple data versions and formats. Moreover, once the candidate resources are identified, she should run an analysis of the data, to better understand their reliability concerning the target task. Additionally, this analysis (see, for instance, the simple data about the number of properties associated with each class) requires a processing phase that is assumed to be set up and run directly by the scientist. As a final step, if the scientist succeeds in finding the data she needs, a transformation process must be run to re-use the relational data in the reference ETR setup. What if the scientist can run all these operations in one single place with the support of ready-to-be-used built-in facilities?</p><p>The idea of LiveSchema precisely arose from this key challenge. Firstly, the gateway aims at supporting scientists in better finding the relational data they need. Indeed, by leveraging the updates of some of the best state-of-the-art catalogs, LiveSchema should offer an aggregation service that allows searching and keeping track of the evolution of the knowledge representation development community in one place.</p><p>Moreover, by implementing some key state-of-the-art libraries, LiveSchema aims to facilitate the data analysis and preparation process. Most of the implemented libraries, indeed, require an ad-hoc set-up and may involve the combination of multiple components and environments, involving some coding and development skills that not all pure data scientists have. In this sense, LiveSchema aims to offer a platform that unites data analysis, data processing, and machine learning model deployment, making them easily accessible, reusable, and less time-consuming.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Data Architecture</head><p>The current version of LiveSchema is grounded in the CKAN<ref type="foot" target="#foot_1">2</ref> open-source data management system which is widely recognized as one of the most reliable tools for managing open data. We concentrate on the fundamental distinction in CKAN which informs the data architecture of LiveSchema, namely that between dataset and resource. <ref type="foot" target="#foot_2">3</ref> A dataset is defined as a set of data (e.g., BBC Sport Ontology) which may contain several resources representing the physical embodiment of the dataset in different downloadable formats (e.g., BBC Sport Ontology in T U R T L E , F C A formats). This distinction allows us, as a major advance from mainstream catalogs such as <ref type="bibr" target="#b12">[13]</ref>, to exploit fine-grained metadata properties from the Application Profile for European Data Portals (DCAT-AP), <ref type="foot" target="#foot_3">4</ref> which makes a conceptually identical distinction between dataset and distribution. The additional advantage of using DCAT-AP is that it organizes metadata into mandatory, recommended, and optional properties which are considered the key for facilitating different levels of semantic interoperability amongst data catalogs.</p><p>We now elucidate the metadata specification, i.e. the selected metadata properties for datasets and distributions considered for the current version of LiveSchema: i. Dataset: Notice that the distinction between dataset and distribution metadata is non-trivial in the sense that metadata properties like format, license, byte size and download url are associated to a distribution and not to the dataset itself.</p><formula xml:id="formula_0">-M A N D A T O R Y : description, title; -R E C O M M E N D E D : dataset distribution,</formula><p>Our first observation concerns the two major advantages that the aforementioned data distinction and metadata specification bring to LiveSchema. Firstly, metadata enforces 'FAIR'ification <ref type="bibr" target="#b13">[14]</ref> of the KG schemas (which are 'data' in this case), thus rendering them findable, accessible, interoperable, and reusable for the machine learning tasks that LiveSchema targets. <ref type="foot" target="#foot_4">5</ref>Secondly, as a consequence of the first advantage, the metadata-enhanced KG schemas also play a pivotal role in initiating, enhancing, and sustaining reproducibility <ref type="bibr" target="#b15">[16]</ref> which is key for LiveSchema vis-à-vis the target machine learning ecosystem in which it participates.</p><p>Our second observation concerns the future extensibility of the metadata specification of LiveSchema. The starting distinction between dataset and distribution can help bootstrap the extension of the initial metadata specification to ontology-specific metadata which, mutadis mutandis, preserves the same distinction via the notions of ontology conceptualization and ontology serialization <ref type="bibr" target="#b16">[17]</ref>. One of the key advantages of using ontology-specific metadata in LiveSchema is that the user can perform a highly customized (conjunctive) search, for instance, even at the level of logical formalism or ontology design language, thus retrieving the most compatible schema for the machine learning task at hand. In this direction, we plan to exploit the MOD <ref type="bibr" target="#b17">[18]</ref> ontology metadata proposals in the immediate future.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Evolving LiveSchema</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Data Collection and Development</head><p>At the current state, LiveSchema relies on four main state-of-the-art catalogs, namely LOV, DERI,<ref type="foot" target="#foot_5">6</ref> FINTO<ref type="foot" target="#foot_6">7</ref> and a custom catalog which is still under construction, <ref type="foot" target="#foot_7">8</ref> where some selected resources are stored.</p><p>Each catalog is associated with a provider, which is the person or organization that uploaded the data set and is in charge of its maintenance in the source catalog. From each catalog, multiple data sets have been individually scraped and uploaded in an automated way. Currently, LOV is the catalog providing most of the data sets, being one of the most widely used catalogs of vocabularies in the semantic web context. The extension of LiveSchema with new catalogs is part of the immediate future work.</p><p>Given the selected catalogs, 26 types of metadata have been scraped, namely: .i Catalog: id, name, title, logo, URL, description; .ii Data-Set: id, name, title, notes (description), issued, modified (last modified), language, uri (landingPage), contact-uri (homepage / contactPoint), maintainer / publisher (Provider), author / creator (Provider), license-id (license), license-url (license), owner-org (Catalog), version (versionInfo), tags (keyword), source; .iii Provider: id, name, title, uri.</p><p>We carefully checked the dataset during data scraping to ensure that LiveSchema is not breaking any license agreement. Currently, five kinds of licenses are admitted given their restrictions (all of them are part of the Creative Commons<ref type="foot" target="#foot_8">9</ref> initiative). These license constraints need to be checked since we both provide access and we manipulate their content to provide the following resources. As we parse them from their source from various sets of formats, <ref type="foot" target="#foot_9">10</ref>we serialize them into the most common ones, namely R D F and T u r t l e . More advanced output formats can be generated through the processing operations enabled by the LiveSchema services, namely C S V (where all the triples and metadata of the input relational data are stored in a datasheet format), C U E (where all the cue metrics are provided), F C A (i.e., the FCA transformation matrix result), V I S (the format that can be used to enable visualization services functionalities), E M B (the format used to generate a statistical model based on a knowledge embedding process).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Stoking LiveSchema</head><p>LiveSchema is managed by a group of knowledge experts, software engineers, and data scientists who contribute to the development and evolution of the whole system. This group of experts, whom we call here LiveSchema administrators, or simply admins, besides handling maintenance issues, are in charge of applying the evolution component functionalities. These functionalities play central roles in the process of populating the catalog. Two evolution operations can be applied by the LiveSchema admin. Firstly, LiveSchema provides an automated evolution process, which is composed of a parsing phase and a scraping phase. Few checkpoints are released for administrators to supervise the output of the automatic processes. Secondly, manual datasets uploading, reviewing, and managing are also available through the usage of LiveSchema services.</p><p>An example of the manually created list, containing new useful data sets, which are not present in the other selected catalogs, with all their relative information and metadata, is accessible. <ref type="foot" target="#foot_10">11</ref> Some of these data sets are not directly obtainable from the web and they had to be downloaded, unzipped, or edited, and then uploaded on GitHub in order to render them collectible using an URL link. Once the data is scraped, a second key semi-automated parsing process is applied.</p><p>The parsing process is very simple, it is executed iteratively and has the goal of producing two main outputs, namely a set of serialized data sets and a set of parsed data sets. The first output is produced by scanning the data sets list and parsing it using RDFlib python library, <ref type="foot" target="#foot_11">12</ref>namely a library that is used to process RDF resources. Here the produced output is used to generate more standard reference formats, which, in the current setting are represented by R D F and T u r t l e . We also allow for the generation of an x l s x (or c s v ) file encoding all the information (e.g., triple and metadata) about the data set to easily enable the other applications provided by the catalog. In this step, the key role of the admin in charge of the parsing process is to edit the data set list in order to filter out undesired data sets and parse only the ones that are required. The second output is produced by scanning each triple of the input data set. The filtering among the predicates to specify the focus of the dataset is applied just before the application of some services. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Forging Datasets</head><p>All the datasets that are gathered from the source catalogs and uploaded to LiveSchema can be then transformed and used as input of the available functionalities. The current LiveSchema version contains six main functionalities, which are 1) FCA generator, namely the process by which data can be converted in the FCA format (FCA); 2) CUEs generator, i.e., the process by which the CUEs (as defined in Section 2) are generated and encoded in the C U E format; 3) Visualization generator, namely the process by which the input data can visualized and analyzed (see V I S format; 4) Knowledge Embedder, i.e., the application by which a model can be created out of the input data, by applying one or some of the libraries provided by the PyKEEN package <ref type="bibr" target="#b18">[19]</ref> (see E M B reference format); <ref type="foot" target="#foot_12">13</ref> 5) the Query Catalog service, which allows running SPARQL queries; 14 6) the knowledge graph visualizer, namely an implementation of the WebVOWL library. <ref type="foot" target="#foot_14">15</ref> This set of functionalities can be easily accessed and reused utilizing APIs services, and can also be easily extended, e.g., 4), 5) and 6) can be run by directly using . r d f files as input. Each functionality may require an ad hoc format to produce the output, and, in some cases, it may have some dependencies with the input format of other functionalities, e.g., 1), 2) and 3) involving new formats.</p><p>Figure <ref type="figure" target="#fig_0">1</ref> above provides an example of the LiveSchema data set management, where all the available functionalities for managing, analyzing, and transforming the data are presented. <ref type="foot" target="#foot_15">16</ref>A set of metadata, tags, and information about the reference source catalog are also provided to users on the top left of the link. All the new formats (if present) are accessible on the corresponding functionality page.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">LiveSchema Components</head><p>The main components of LiveSchema are combined as in Figure <ref type="figure" target="#fig_1">2</ref>. LiveSchema includes five main components (See Figure <ref type="figure" target="#fig_1">2</ref>): (1) user interfaces UIs; (2) the APIs, provided by the CKAN platform, and that we partially customized according to our set-up needs, provide the main accesses to the LiveSchema environment; (3) stoking components; (4) forging components cover the main novel contributions of the LiveSchema initiative and offer the possibility to harvest, generate and process data; (5) Storage allows to collect in one place all the data collected by other catalogs, provided by the users or generated through the services. All these components are grouped into three main layers: the presentation layer, the service layer, and the data layer. Presentation layer. This layer enables a community of users: .i to maintain the whole gateway and its applications, and .ii to suggest and upload new resources or edit some already existing resources. LiveSchema is mainly managed by a group of expert knowledge engineers, software engineers, and data scientists who contribute to the development and evolution of the whole catalog. Moreover, a group of guest users can also be involved in the collaborative development of the storage, by uploading and editing some new data sets and, possibly, creating new input reference catalogs, following well-founded guidelines provided by the knowledge engineers that administrate the ecosystem. The definition of the types of access and the different roles played by the LiveSchema users is part of the immediate future work. The APIs allow the users to exploit all of the website's core functionalities by external code. Using the API, developers will be able to: get JSON-formatted lists of vocabularies, with providers (namely, the agent who created the data set), source catalogs, or other LiveSchema information; get a full JSON representation of a data set or other related information derived from the analysis of the data set; search for data sets, providers, or other resources matching a query; create, update and delete data sets, with related metadata and information; get an activity stream of recently changed data set on LiveSchema, obtaining also the versioning information of each resource The UIs allow users to access data and functionalities. Two types of user interfaces are present, namely front-end and back-end user interfaces. The former is a customization of the standard CKAN template, where the home page allows to access all the contents of the website through five main widgets: 1) a menu with the top-level categories of the catalog, 2) a search form to easily access and browse data sets, 3) a showcase of the top services and a list of the source reference catalogs, 4) recent activities. Differently, the back-end user interface can be accessed only with credentials and allows for the editing and submission of existing or new data, or it enables the usage of some more applications.</p><p>Service layer. The stoking components are mainly necessary to check and gather any new knowledge resources from a set of previously selected catalogs. LiveSchema mainly relies on manual processes and a semi-automated process for data insertion. The former can be applied by any type of user, by submitting a new resource through a dedicated panel, but requires a review process from the administrator users. The latter is applied by selecting source catalogs as input and can be used to keep track of their updates. This stoking facility can be primarily customized by determining how many times the source catalogs must be checked and by defining what types of data sets can be collected and uploaded into the main storage. Currently, the quality criteria to allow the uploading of a data set, is the size, the type of license, and the correct format of its content. Along with the stoking components, the LiveSchema forging components encode a set of functionalities, which are aimed at the analysis and transformation of data, and the generation of new formats. All these functionalities are aimed at supporting scientists in the re-usage of the selected relational data.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Using LiveSchema</head><p>The scope of this section is mainly to show how the LiveSchema processing component works. Through a running example, we illustrate how a user can exploit main functionalities. All the described operations can be directly tested by exploring and using the LiveSchema ecosystem, which is accessible at http://liveschema.eu/. The admin functionalities can be accessed and tested at http://liveschema.eu/user/login, by using ' r e v i e w e r ' as admin/password.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.1.">Analyzing Relational Data</head><p>As an example scenario, suppose we need to run a standard entity type recognition task, as it is described in <ref type="bibr" target="#b7">[8]</ref> and <ref type="bibr" target="#b11">[12]</ref>, where we may need to recognize objects of the type 'Person' across a set of multiple tabular data, coming, for instance, from an open data repository. This may involve the need to find a reference relational model with .i the target class and corresponding label; .ii possibly a huge number of properties for the target class, in order to increase the probability to match some of the input test properties; .iii possibly a low number of overlapping properties, in order to decrease the number of false negative/positive.</p><p>A LiveSchema user can perform a simple search across the available data sets that are present in the catalog and then run an analysis to select the best. The LiveSchema search facility exploits the CKAN search engine and allows for a quick 'Google-style' keyword search. All the data sets, providers, and group fields are searchable and the users can use all of them to research the desired entity. Thanks to this search functionality it is possible to provide a complete and customized service to the scientist looking for the desired ontology. The basic supported search options are .i search over all the data sets attributes, namely by using any of the applied metadata; .ii full-text search; .iii fuzzy-matching, namely an option to search for closely matching terms instead of exact matches; .iv search via API. Now, suppose that the user identifies three candidate resources for the goal ETR task, namely Schema.org <ref type="foot" target="#foot_16">17</ref> (reference standard to support the indexing of web documents by Google<ref type="foot" target="#foot_17">18</ref> ), FOAF <ref type="foot" target="#foot_18">19</ref> (a widely used vocabulary in the context of social networks) and the BBC sport ontology <ref type="foot" target="#foot_19">20</ref> (the ontology used by the BBC to model supports events, and roles). The next step is to access each single data set and check its meta-information, which can be done by first generating the F C A format for the selected resources.</p><p>Each LiveSchema data set has a dedicated page collecting its information, and where each processing functionality can be accessed. Here information about the related source catalog is provided as well, and the available standard reference formats can be downloaded. The FCA functionality can be accessed through the corresponding tab and allows for the generation of the corresponding matrix for each given input relational model. On the FCA service page is also possible to customize the generation of the matrix by filtering the target predicates. Then, multiple insights can be extracted by using the functionalities represented by each tab on the data set page. By downloading all the cue information comparisons between the three representations of 'Person', provided by each ontology can be run. Figure <ref type="figure" target="#fig_2">3</ref> (a) represents the cue values for the given resources. From a quick benchmark is clear that in Schema.org, even if the cue of Person is not at the top, the given class has a high centrality with a score of 23.</p><p>Besides the quantification of the cues, further analysis can be run by visualizing the intersection of some of the top classes of the given resources. Figure <ref type="figure" target="#fig_2">3</ref> (b) represents an example of knowledge lotus that can be extracted by the input resources. Knowledge lotuses are venn diagrams that can be used to focus on specific parts of the input resources and they are particularly useful to represent the diversity of classes in terms of their (un-)shared properties. The yellow petals of the lotus show the number of properties that are distinctive for the given class. In  Further analysis can be run by applying the UpSet (multiple set) visualization facilities, which allows us to analyze the intersections between classes, by selecting more than 6 sets (the limit for knowledge lotuses). LiveSchema allows for both knowledge lotuses and UpSet visualization by embedding the functionalities of the intervention visualization environment. <ref type="foot" target="#foot_20">21</ref>This environment was created for the visualization of multiple genomic regions and gene sets (or lists of items). The main goal of the provided visualization options is to facilitate the analysis and interpretation of the input resource. An illustrative example of the representation of a resource utilizing the UpSet module is provided by  the size of the properties intersection set.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.2.">Embedding Relational Data</head><p>Once the scientist has selected her resource, she is ready to embed it and generate a statistical model out of it. Notice that, in the current release of LiveSchema we allow for distributional embedding techniques only. The implementation of symbolic approaches, such as Inductive Logic Programming for addressing new tasks like, for instance, class expression learning, is part of the immediate future work. In this current setting, LiveSchema relies on a recent library collecting most of the state-of-the-art techniques for graph embedding, namely the PyKEEN library <ref type="bibr" target="#b18">[19]</ref>. PyKEEN is a widely used solution for generating custom embedding models. It allows selection across a wide range of training approaches with multiple parameters and will output a . p k l file which can be directly imported inside ML pipelines.</p><p>Figure <ref type="figure" target="#fig_5">5</ref> demonstrates a screenshot of the LiveSchema KnowledgeEmbedder page, various parameters can be selected to obtain the specific learning goal. We can select the "embedding model" where we can select state-of-the-art algorithms, like TransE, RESCAL or DistMult <ref type="bibr" target="#b19">[20]</ref>; and settings like the "loss function", which is typically used to minimize the error of the model and can be used for reducing multiple features of the models to a single number, namely a scalar value, which allows candidate solutions to be ranked and compared <ref type="bibr" target="#b20">[21]</ref>.</p><p>Notice that in LiveSchema we have data sets encoding relational models with no instance data (e.g., we have the DBpedia schema, but we do not have the so-called ABOX). This did not prevent us to adapt the embedding process and focus on the schema level only (relying on relational data we always have, indeed, triples: heads, tails, and relations). This, besides opening the possibility to test a new application scenario, does not exclude the possibility of applying the standard approach where populated schemas are used as input. The population of LiveSchema with this kind of data is part of the immediate future work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Discussion</head><p>We believe that LiveSchema could be a useful support to study relational data for both knowledge representation tasks (e.g., designing a knowledge base for enabling the interoperability between systems), and machine learning tasks, in particular the ones that rely on relational data structures for training their models. In this section, we discuss the implications of the initiative. Moreover, by identifying the limitations of our current setting, we also discuss opportunities for future work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.1.">Implications</head><p>Firstly, LiveSchema can support scientists in finding the relational data they need. By leveraging the updates of some of the best state-of-the-art catalogs, LiveSchema can offer an aggregation service that allows keeping track of the evolution of the knowledge representation development community in one place. Notice that this does not aim to substitute the function of each single source catalog. The scientist can indeed access the source catalog and related ad hoc services, if needed, directly from the LiveSchema gateway, this being also an opportunity to increase the visibility of the vocabularies themselves.</p><p>Another key point is that LiveSchema can represent an opportunity to bridge the gap between two key artificial intelligence communities, namely the knowledge representation and the machine learning community. While most of the data that are present in LiveSchema are indeed in a format that is compliant with the knowledge representation applications requirements, each data set can be also transformed so that it can be easily employed in machine learning set-ups. The analysis and embedding facilities offer further support in this direction. We believe that this is a way of supporting the exploitation of the huge amount of work done by the community and of making the relational model more accessible to machine learning scientists.</p><p>Moreover, we implement state-of-the-art libraries to support data scientists in the data analysis and preparation phases. Most of the implemented libraries require coding and development skills, which will limit the usage of data scientists. To solve this issue, LiveSchema offers a platform that unites data analysis, data preprocessing, and machine learning model deployment, which makes them easily accessible and usable.</p><p>Finally, the overall project was also devised to pave the way for large case studies. Integrating knowledge representation and machine learning scenarios may indeed be devised in a different way of designing relational structures, with a different focus on some of their features or constraints (e.g., the number of properties to be used for describing a class or the overlapping between classes). Moreover, data scientists, reusing relational models for their predictive tasks, may better realize what relational models can be better than others about the specific learning target, and how they should be tuned to better support their task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.2.">Limitations</head><p>Developing LiveSchema as a community of data scientists that exchange and reuse data to the benefit of the AI community is our long-term objective and this triggers the agenda for immediate future work. To achieve this goal, with the current set-up, there is still a gap that needs to be bridged.</p><p>As long as the LiveSchema observatory will grow, serious challenges about the scalability of the approach still need to be handled. One issue is that through the current version of the evolution component is not possible to automatically check duplicated resources coming from different vocabularies. Another pending issue, which is part of the agenda for the immediate future work, is the definition of a processing component functionality that enables users to work with multiple data sets together, this would be an important option, especially for supporting data integration tasks and the evolution of more robust machine learning models. A possible way of implementing this functionality will be to develop a new version of the current FCA conversion process, where multiple data sets can be given as input and then merged, by computing their similarities, into one single file. The output file will be used as a single resource containing the information of all its component datasets.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="8.">Conclusion</head><p>In this paper, we introduced the first version of the LiveSchema gateway, which aims at exploiting the relational representation of ontologies not only for their classical application but also for their use in machine learning scenarios. The long-term goal of LiveSchema is to leverage the gold mine of data collected by many of the existing relational knowledge representations catalogs and offer a family of services to easily access, analyze, transform, and re-use data, with an emphasis on relational machine learning pipelines setup and predictive tasks.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: LiveSchema data set page (from an admin perspective).</figDesc><graphic coords="6,151.80,161.43,291.68,266.96" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Overview of the LiveSchema components.</figDesc><graphic coords="7,172.63,385.11,250.02,137.83" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: Cue values for the class 'Person'.</figDesc><graphic coords="10,151.80,256.37,291.69,179.95" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: Etypes properties intersection: the UpSet visualization</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 4 .</head><label>4</label><figDesc>Here 8 classes are selected. The blue bars on the left show the size of the classes in terms of the number of properties. The black dots identify the intersections between the classes and the red bars on top of the figure show</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: The Knowledge Embedding interface in Liveschema (zoom in for more details).</figDesc><graphic coords="11,130.96,84.19,333.34,166.02" type="bitmap" /></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">Relational data was proven to be key also in a transfer learning setting<ref type="bibr" target="#b9">[10,</ref><ref type="bibr" target="#b10">11]</ref>.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">https://docs.ckan.org/en/2.9/user-guide.html#what-is-ckan</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="3" xml:id="foot_2">https://docs.ckan.org/en/538-package-install-docs/publishing-datasets.html</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="4" xml:id="foot_3">https://ec.europa.eu/isa2/solutions/dcat-application-profile-data-portals-europe_en/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="5" xml:id="foot_4">To get an idea of what a FAIR catalog consists of, the work presented in<ref type="bibr" target="#b14">[15]</ref> and accessible at https://github.com/ OntoUML/ontouml-models provides a key reference</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="6" xml:id="foot_5">http://vocab.deri.ie/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="7" xml:id="foot_6">http://finto.fi/en/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="8" xml:id="foot_7">http://liveschema.eu/organization/knowdive</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="9" xml:id="foot_8">https://creativecommons.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="10" xml:id="foot_9">https://rdflib.readthedocs.io/en/stable/plugin_parsers.html</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="11" xml:id="foot_10">https://github.com/knowdive/resources/blob/master/otherVocabs.xlsx</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="12" xml:id="foot_11">https://github.com/RDFLib/rdflib</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="13" xml:id="foot_12">https://pypi.org/project/pykeen/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="14" xml:id="foot_13">https://rdflib.readthedocs.io/en/stable/intro_to_sparql.html</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="15" xml:id="foot_14">http://vowl.visualdataweb.org/webvowl.html</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="16" xml:id="foot_15">three current reference resources . r d f , . t t l and . c s v are ready to be downloaded.</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="17" xml:id="foot_16">https://schema.org/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="18" xml:id="foot_17">https://www.google.com/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="19" xml:id="foot_18">http://xmlns.com/foaf/spec/</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="20" xml:id="foot_19">https://www.bbc.co.uk/ontologies/sport</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="21" xml:id="foot_20">https://intervene.shinyapps.io/intervene/</note>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>The research conducted by Mattia Fumagalli is supported by the "Dense and Deep Geographic Virtual Knowledge Graphs for Visual Analysis -D2G2" project, funded by the Autonomous Province of Bolzano. The research conducted by Fausto Giunchiglia and Mayukh Bagchi has received funding from JIDEP -under grant agreement number 101058732. The research conducted by Daqian Shi has received funding from the program of China Scholarships Council (No. 202007820024).</p></div>
			</div>

			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Garbage in, garbage out? do machine learning application papers in social computing report where humanlabeled training data comes from?</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">S</forename><surname>Geiger</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Dai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Qiu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Huang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency</title>
				<meeting>the 2020 Conference on Fairness, Accountability, and Transparency</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="325" to="336" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">Pattern recognition and machine learning</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Anzai</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2012">2012</date>
			<publisher>Elsevier</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title level="m" type="main">The elements of statistical learning: data mining, inference, and prediction</title>
		<author>
			<persName><forename type="first">T</forename><surname>Hastie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Tibshirani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Friedman</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2009">2009</date>
			<publisher>Springer Science &amp; Business Media</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">What is a knowledge graph?</title>
		<author>
			<persName><forename type="first">M</forename></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Domain-Specific Knowledge Graph Construction</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="1" to="7" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">A review of relational machine learning for knowledge graphs</title>
		<author>
			<persName><forename type="first">M</forename><surname>Nickel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Murphy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Tresp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Gabrilovich</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Proceedings of the IEEE</title>
		<imprint>
			<biblScope unit="volume">104</biblScope>
			<biblScope unit="page" from="11" to="33" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Concepts as (recognition) abilities</title>
		<author>
			<persName><forename type="first">F</forename><surname>Giunchiglia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fumagalli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">FOIS</title>
		<imprint>
			<biblScope unit="page" from="153" to="166" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Teleologies: Objects, actions and functions</title>
		<author>
			<persName><forename type="first">F</forename><surname>Giunchiglia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fumagalli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International conference on conceptual modeling</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="520" to="534" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Entity type recognition-dealing with the diversity of knowledge</title>
		<author>
			<persName><forename type="first">F</forename><surname>Giunchiglia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fumagalli</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Seventeenth International Conference on Principles of Knowledge Representation and Reasoning</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">On knowledge diversity</title>
		<author>
			<persName><forename type="first">F</forename><surname>Giunchiglia</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fumagalli</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2019">2019</date>
			<publisher>JOWO</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Ontology-driven cross-domain transfer learning</title>
		<author>
			<persName><forename type="first">M</forename><surname>Fumagalli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Bella</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Conti</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Giunchiglia</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="volume">330</biblScope>
			<biblScope unit="page">249</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Towards understanding classification and identification</title>
		<author>
			<persName><forename type="first">M</forename><surname>Fumagalli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Bella</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Giunchiglia</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Pacific Rim International Conference on Artificial Intelligence</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="71" to="84" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Entity type recognition for heterogeneous semantic graphs</title>
		<author>
			<persName><forename type="first">J</forename><surname>Sleeman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Finin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Joshi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">AI Magazine</title>
		<imprint>
			<biblScope unit="volume">36</biblScope>
			<biblScope unit="page" from="75" to="86" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Linked open vocabularies (lov): a gateway to reusable semantic vocabularies on the web</title>
		<author>
			<persName><forename type="first">P.-Y</forename><surname>Vandenbussche</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Atemezing</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Poveda-Villalón</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Vatant</surname></persName>
		</author>
		<idno type="DOI">10.3233/SW-160213</idno>
	</analytic>
	<monogr>
		<title level="j">Semantic Web Journal</title>
		<imprint>
			<biblScope unit="volume">8</biblScope>
			<biblScope unit="page" from="437" to="452" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">The fair guiding principles for scientific data management and stewardship</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">D</forename><surname>Wilkinson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Dumontier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">J</forename><surname>Aalbersberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Appleton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Axton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Baak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Blomberg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J.-W</forename><surname>Boiten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">B</forename><surname>Da Silva Santos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">E</forename><surname>Bourne</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Scientific data</title>
		<imprint>
			<biblScope unit="volume">3</biblScope>
			<biblScope unit="page" from="1" to="9" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">A fair model catalog for ontology-driven conceptual modeling research</title>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">P F</forename><surname>Barcelos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">P</forename><surname>Sales</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fumagalli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">M</forename><surname>Fonseca</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><forename type="middle">V</forename><surname>Sousa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Romanenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kritz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Guizzardi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Conceptual Modeling: 41st International Conference, ER 2022</title>
				<meeting><address><addrLine>Hyderabad, India</addrLine></address></meeting>
		<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2022">October 17-20, 2022. 2022</date>
			<biblScope unit="page" from="3" to="17" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">The role of metadata in reproducible computational research</title>
		<author>
			<persName><forename type="first">J</forename><surname>Leipzig</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Nüst</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">T</forename><surname>Hoyt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Ram</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Greenberg</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Patterns</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="page">100322</biblScope>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Omv-ontology metadata vocabulary</title>
		<author>
			<persName><forename type="first">J</forename><surname>Hartmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Sure</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Haase</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Palma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Suarez-Figueroa</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ISWC</title>
				<imprint>
			<publisher>Citeseer</publisher>
			<date type="published" when="2005">2005</date>
			<biblScope unit="volume">3729</biblScope>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Mod: metadata for ontology description and publication</title>
		<author>
			<persName><forename type="first">B</forename><surname>Dutta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Nandini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">K</forename><surname>Shahi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Dublin Core and Metadata Applications</title>
				<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="1" to="9" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Pykeen 1.0: a python library for training and evaluating knowledge graph embeddings</title>
		<author>
			<persName><forename type="first">M</forename><surname>Ali</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Berrendorf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">T</forename><surname>Hoyt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Vermue</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sharifzadeh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Tresp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lehmann</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="page" from="3723" to="3728" />
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Knowledge graph embedding: A survey of approaches and applications</title>
		<author>
			<persName><forename type="first">Q</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Mao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Guo</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Knowledge and Data Engineering</title>
		<imprint>
			<biblScope unit="volume">29</biblScope>
			<biblScope unit="page" from="2724" to="2743" />
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Deep learning in neural networks: An overview</title>
		<author>
			<persName><forename type="first">J</forename><surname>Schmidhuber</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Neural networks</title>
		<imprint>
			<biblScope unit="volume">61</biblScope>
			<biblScope unit="page" from="85" to="117" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
