<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">EmEL ++ : Embeddings for  ++ Description Logic</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Sutapa</forename><surname>Mondal</surname></persName>
							<email>sutapa.mondal@tcs.com</email>
							<affiliation key="aff0">
								<orgName type="laboratory">Knowledgeable Computing and Reasoning (KRaCR) Lab</orgName>
								<orgName type="institution">IIIT-Delhi</orgName>
								<address>
									<country key="IN">India</country>
								</address>
							</affiliation>
							<affiliation key="aff2">
								<orgName type="institution">TCS Research &amp; Innovation</orgName>
								<address>
									<country key="IN">India</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Sumit</forename><surname>Bhatia</surname></persName>
							<email>sumitbhatia@in.ibm.com</email>
							<affiliation key="aff1">
								<orgName type="institution">IBM Research AI</orgName>
								<address>
									<settlement>New Delhi</settlement>
									<country key="IN">India</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Raghava</forename><surname>Mutharaju</surname></persName>
							<email>raghava.mutharaju@iiitd.ac.in</email>
							<affiliation key="aff0">
								<orgName type="laboratory">Knowledgeable Computing and Reasoning (KRaCR) Lab</orgName>
								<orgName type="institution">IIIT-Delhi</orgName>
								<address>
									<country key="IN">India</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff3">
								<orgName type="institution">Stanford University</orgName>
								<address>
									<addrLine>March 22-24</addrLine>
									<postCode>2021</postCode>
									<settlement>Palo Alto</settlement>
									<region>California</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">EmEL ++ : Embeddings for  ++ Description Logic</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">F58ED92114430FA262270EFF9DE3104A</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T08:17+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Ontology</term>
					<term> ++</term>
					<term>Description Logic</term>
					<term>Geometric Embeddings</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Knowledge graph (KG) embedding models have recently gained increased attention. However, most of the existing models for KG embeddings ignore the structure and characteristics of the underlying ontology. In this work, we present EmEL ++ embeddings -an ontology-based embedding model for the  ++ description logic. EmEL ++ maps the classes and the relations in an ontology to an n-dimensional vector space such that the relations between classes and relations in the ontology are preserved in the vector space. We evaluate the proposed embeddings on four different datasets and show that the proposed embeddings outperform the traditional knowledge graph embeddings on the subsumption reasoning task.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Methods for learning embedding functions that map the underlying entities to a vector space have gained significant attention in recent times. Different methods for learning embedding functions try to preserve the critical properties of, and relations between, the underlying entities (such as words, concepts, documents, nodes and edges in a graph) in the 𝑛-dimensional vector space. A variety of embeddings for Knowledge Graphs <ref type="bibr" target="#b0">[1,</ref><ref type="bibr" target="#b1">2,</ref><ref type="bibr" target="#b2">3,</ref><ref type="bibr" target="#b3">4,</ref><ref type="bibr" target="#b4">5,</ref><ref type="bibr" target="#b5">6]</ref> have been proposed. These methods differ in terms of the underlying properties of the knowledge bases preserved in the vector space. For example, the TransE <ref type="bibr" target="#b0">[1]</ref> model considers the relations in a KG as a translation operator over the entities, TransH <ref type="bibr" target="#b1">[2]</ref> models the relations as a hyperplane in the vector space to allow for reflexive, one-to-many, many-to-one, and many-to-many relations, DistMult <ref type="bibr" target="#b2">[3]</ref> uses Matrix Factorization based approach to bring similar entities and relations together in the vector space. However, most of the KG embedding models focus on capturing the structural properties of the graph and the interaction between the entities and do not take into account the constraints and characteristics of the underlying ontology. Consequently, the embeddings produced by such methods are not suited for reasoning tasks such as classification, satisfiability and consistency checking. Kulmanov et al. <ref type="bibr" target="#b6">[7]</ref> have recently proposed EL Embeddings (ElEm) and have described a geometric interpretation of embedding  ++ ontologies in an 𝑛-dimensional vector space (Section 3). However, the ElEm embeddings are limited in terms of the coverage of  ++ constructs as they ignore the role oriented constructs in such ontologies. Further, the use case considered by them is predicting protein-protein interactions that is modeled as a traditional link prediction task in knowledge bases. In this work, we address the limitations of ElEm embeddings and propose EmEL++ embeddings. We build upon the framework introduced by Kulmanov et al. <ref type="bibr" target="#b6">[7]</ref> and extend it to offer more complete coverage of the  ++ semantics (Section 3.1) for performing subsumption reasoning task. Baader et al. <ref type="bibr" target="#b7">[8]</ref> have shown that all the standard reasoning tasks in  ++ ontologies can be reduced to subsumption task. Thus, to the best of our knowledge, we present the first attempts at performing reasoning tasks by embedding ontologies in a vector space. We compare our proposed approach using four ontologies of different sizes and characteristics. Our evaluation shows that EmEL ++ outperform the traditional KG embeddings at the reasoning task and are also able to preserve the characteristics of underlying ontologies better. We would also like to emphasize that performing reasoning in the vector space is critical as it has the potential to speed up the reasoning process significantly. As we describe in Section 4, the subsumption task in vector space involves computing distances between the source class and all the other classes in the ontology. In the worst case, this is an 𝑂(𝑛) operation where 𝑛 is the number of classes. Thus, irrespective of the complexity of the underlying ontology, the subsumption task could be performed in 𝑂(𝑛) time. Further, with uses of techniques such as semantic hashing or binarized embeddings <ref type="bibr" target="#b8">[9]</ref>, the similarity based search operations can be performed in 𝑂(1) time. Therefore, we believe that embedding based approaches, despite their lower accuracies than standard reasoners and no theoretical guarantees of performance, offer a promising direction of future research to develop more efficient reasoners, especially for more complex description logics such as  (OWL 2 DL).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head><p>A wide range of methods for computing KG embeddings have been proposed. Node2Vec <ref type="bibr" target="#b9">[10]</ref> initiated the idea of learning features for networks addressing the scalability challenge. Although their results are decent for link prediction task but they make assumptions on conditional independence of the feature space which may not hold true in real world scenarios. Eventually, this concept got popularized with KGs wherein, a fact is represented as a triple of the form (h,r,t). Semantic matching KG models exploit similarity based scoring functions and match latent semantics of entities and relations based on vector representations. For example, <ref type="bibr" target="#b3">[4]</ref> proposed RESCAL which uses multiple matrices to represent relations among entities but scalability remains an issue. <ref type="bibr" target="#b2">[3]</ref> proposed DistMult to overcome challenges of RESCAL. Although, DistMult is similar to RESCAL but DistMult ensures low number of parameters for relations by restricting the matrices. Later, translational based models for KG embeddings used distance based scoring functions. This technique gained most attention due to its simplicity to measure the correctness of a fact. It measures the plausibility of a fact as distance between the entities after translation carried out by relation. TransE <ref type="bibr" target="#b0">[1]</ref>, TransH <ref type="bibr" target="#b1">[2]</ref> and TransR <ref type="bibr" target="#b5">[6]</ref> being different variants of the same. TransE being the most representative model, treats the relations as translations such that given a fact, relation vector r minimizes the distance between h and t in vector space. Whereas, TransH interprets relations as translating operator on a hyperplane and TransR models entities and relations in two distinct spaces, performing translation on corresponding relation space. <ref type="bibr" target="#b10">[11]</ref> worked on a novel approach inspired by the theory of quantum logic to embed a Knowledge Base (KB) in  description logic. Existing works on ontology embedding such as Onto2vec <ref type="bibr" target="#b11">[12]</ref> focused on using word2vec as an underlying model. Most of this work focuses on encoding the entities and relations, but they lack in handling the complex relations in an ontology. Recently, <ref type="bibr" target="#b12">[13]</ref> pointed out the lack of expressivity in the classical approaches to model relations for KG embeddings. Moreover, their work indicates that geometric models are a better way to learn embeddings for ontologies. Further, <ref type="bibr" target="#b13">[14]</ref> present a new approach using deep learning with knowledge based systems to emulate reasoning structure. Although, they perform experiments on a synthetic and a non-synthetic dataset, but in our work we look at multiple datasets with different characteristics and reason about their varying performances. However, <ref type="bibr" target="#b6">[7]</ref> tries to overcome the drawbacks of KG embeddings but it does not address all the  ++ constructs that are relevant to capture the relations present in an ontology. Further, the evaluation is focused only on link prediction task.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Background</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Embedding  ++ Ontologies in a Vector Space</head><p>A description logic signature is a tuple ⟨𝑁 𝐶 , 𝑁 𝑅 , 𝑁 𝐼 ⟩, where 𝑁 𝐶 , 𝑁 𝑅 , and 𝑁 𝐼 are countably infinite, mutually disjoint sets of concept names, role names, and individual names respectively. In the following discussion, {𝐶, 𝐶 1 , 𝐶 2 , 𝐷, 𝐸, ⊤, ⊥} ∈ 𝑁 𝐶 , {𝑅, 𝑆, 𝑅 1 , 𝑅 2 } ∈ 𝑁 𝑅 , and {𝑎, 𝑏} ∈ 𝑁 𝐼 .</p><p>All the axioms in the  ++ description logic can be reduced to one of the normal forms <ref type="bibr" target="#b14">[15]</ref> as follows:</p><p>1. All the TBox axioms are in one of the four normal forms (𝐶 1 ⊑ 𝐷, 𝐶 1 ⊓ 𝐶 2 ⊑ 𝐷, ∃𝑅.𝐶 1 ⊑ 𝐷, and 𝐶 1 ⊑ ∃𝑅.𝐶 2 ). 2. The bottom concept can only appear on the right side of the equations, and can only appear in the first three normal forms. 3. All role inclusions are of the form 𝑅 ⊑ 𝑆 or 𝑅 1 •𝑅 2 ⊑ 𝑆.</p><p>Further, the instantiation and role assertion axioms in the ABox can be converted into TBox axioms as follows:</p><formula xml:id="formula_0">𝐶(𝑎) ⟶ {𝑎} ⊑ 𝐶 𝑅(𝑎, 𝑏) ⟶ {𝑎} ⊑ ∃𝑅.{𝑏}</formula><p>Thus, with the above transformations, all the axioms in an  ++ ontology can be reduced to one of the normalized forms and the task of embedding ontologies in a vector space requires us to learn mapping functions for classes and relations that are part of the normal forms. Typically, for mapping the entities of interest to a vector space, requires to learn a mapping function subject to certain constraints, encoded in the form of an objective function that is optimized during the training phase. These objective functions are designed such that specific properties of the underlying entities are also retained in the vector space. For example, the word2vec <ref type="bibr" target="#b15">[16]</ref> model for word embeddings minimizes the distance between contextually similar words, RDF2Vec <ref type="bibr" target="#b16">[17]</ref> adapts language modeling approach to capture local information from the graph sub-structures, and TransE <ref type="bibr" target="#b0">[1]</ref> model for KGs. Similarly, in this work we learn mapping functions that can embed  ++ ontologies in a vector space. In order to do so, we build upon and extend the framework proposed by Kulmanov et al. <ref type="bibr" target="#b6">[7]</ref> that interprets a class in the ontology as an 𝑛-ball (defined by its radius and center) in the vector space. Let us consider two classes 𝐶 and 𝐷 such that 𝐶 ⊑ 𝐷. Let these two classes be represented by their respective 𝑛-balls 𝑏 𝑐 and 𝑏 𝑑 in the vector space such that 𝑏 𝑐 ∶ {𝑐 ⃗, 𝑟 𝑐 } and 𝑏 𝑑 ∶ {𝑑 ⃗ , 𝑟 𝑑 }; where 𝑐 ⃗ and 𝑑 ⃗ are the centers and 𝑟 𝑐 and 𝑟 𝑑 are the radii of the respective 𝑛-balls. Geometrically, if 𝐶 ⊑ 𝐷, the mapping function should aim to ensure that the 𝑏 𝑐 lies inside 𝑏 𝑑 (Figure1 (a)). Similarly, if 𝐶 and 𝐷 are disjoint, the respective 𝑛-balls should not overlap with each other in the vector space (Figure1 (b)). Further, similar to the TransE model <ref type="bibr" target="#b0">[1]</ref>, the relations in the ontology are interpreted as translations operating on the classes. More specifically, if 𝐶 ⊑ ∃𝑅.𝐷, the center of 𝑛-ball representing 𝐶 can be moved to the center of the 𝑛-ball representing 𝐷 (Figure1 (c)).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Loss Functions</head><p>With the intuitive framework described above, let us now describe the objective functions that should be optimized during the training phase to learn the mapping functions. Let 𝑒 𝑣 ∶ 𝐂 ∪ 𝐑 ⟼ 𝑅 𝑛 be the mapping function that maps each class and relation to a unique vector in the 𝑛-dimensional embedding space. For 𝐶 𝑖 ∈ 𝐂, the resulting vector corresponds to the centre of the 𝑛-ball representing the class. Further, let 𝑒 𝑟 ∶ 𝐶 ⟼ 𝑅 + be the mapping function that maps each class into a non-negative real number, that represents the radius of the 𝑛ball corresponding to class 𝐶. Thus, the pair (𝑒 𝑣 , 𝑒 𝑟 ) of functions represents the operations needed to embed an  ++ ontology into an 𝑛-dimensional space. We now describe the various loss functions to represent the different constructs in  ++ . The total loss that needs to be minimized during the learning process is the sum of the individual loss functions.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.1.">Loss Functions for the Four Normal Forms:</head><p>As described before, the first normal form (𝐶 ⊑ 𝐷) when embedded in a vector space can be interpreted geometrically as two 𝑛-balls, such that the 𝑛-ball corresponding to class 𝐶 lies inside the 𝑛-ball corresponding to 𝐷. Hence, our mapping functions 𝑒 𝑣 and 𝑒 𝑟 should bring the centers of the two classes closer to each other, and give the sub-class a smaller radius than the super-class. The loss function presented in Equation 1 captures this intuition and penalizes the mappings that do not adhere to these constraints. Also note that in addition to the above constraints, we also add margin loss (𝛾 ) and a normalization loss that brings the centres of 𝑛-balls of all the classes on the unity sphere. </p><formula xml:id="formula_1"> 𝐶⊑𝐷 (𝑐, 𝑑) = max ⎛ ⎜ ⎜ ⎜ ⎜ ⎜ ⎝ 0, ( ‖ ‖</formula><formula xml:id="formula_2">−𝛾 ) ⎞ ⎟ ⎟ ⎟ ⎟ ⎟ ⎠ + | | | ‖ ‖ 𝑒 𝑣 (𝑐) ‖ ‖ − 1 | | | + | | | ‖ ‖ 𝑒 𝑣 (𝑑) ‖ ‖ − 1 | | |<label>(1)</label></formula><p>In the vector space, the second normal form, i.e., 𝐶 ⊓ 𝐷 ⊑ 𝐸, implies that the 𝑛-ball for class 𝐸 should completely engulf the area of intersection of 𝑛-balls for classes 𝐶 and 𝐷. The first term in the loss function (Equation <ref type="formula" target="#formula_3">2</ref>) imposes a penalty if the classes 𝐶 and 𝐷 are disjoint. The second and third terms together enforce that the center of the 𝑛-ball for class 𝐸 lies in the area of intersection of 𝑛-balls for classes 𝐶 and 𝐷 such that it satisfies the normal form. Finally, the fourth term requires the radius of the 𝑛-ball of 𝐸 to be greater than the smallest radii among 𝑛-balls of 𝐶 and 𝐷. </p><formula xml:id="formula_3">+ | | | ‖ ‖ 𝑒 𝑣 (𝑐) − 1 ‖ ‖ | | | + | | | ‖ ‖ 𝑒 𝑣 (𝑑) − 1 ‖ ‖ | | | + | | | ‖ ‖ 𝑒 𝑣 (𝑒) − 1 ‖ ‖ | | |<label>(2)</label></formula><p>The first two normal forms are concerned with the mappings of classes and properties of their respective 𝑛-balls in the vector space. The next two normal forms, i.e., NF3 and NF4, involve relations and how they are associated with the classes. Recall that similar to TransE <ref type="bibr" target="#b0">[1]</ref>, we consider the relations in ontology as translations that operate on class instances. Consider the normal form 𝐶 ⊑ ∃𝑅.𝐷. In the vector space, 𝐶 and 𝐷 are represented as two 𝑛-balls 𝑏 𝑐 and 𝑏 𝑑 , respectively. If 𝑒 𝑣 (𝑅) is the vector for 𝑅 in vector space, then adding 𝑒 𝑣 (𝑅) to a point in 𝑏 𝑐 should move it to a point in 𝑏 𝑑 (i.e., 𝑅 translates the points in 𝑏 𝑐 to points in 𝑏 𝑑 ). The following loss functions capture these semantics as expressed by the third and fourth normal forms.</p><formula xml:id="formula_4"> 𝐶⊑∃𝑅.𝐷 (𝑐, 𝑑, 𝑟) = max ( 0, ( ‖ ‖ 𝑒 𝑣 (𝑐) + 𝑒 𝑣 (𝑟) − 𝑒 𝑣 (𝑑) ‖ ‖ + 𝑒 𝑟 (𝑐) − 𝑒 𝑟 (𝑑) − 𝛾 ) ) + | | | ‖ ‖ 𝑒 𝑣 (𝑐) ‖ ‖ − 1 | | | + | | | ‖ ‖ 𝑒 𝑣 (𝑑) ‖ ‖ − 1 | | | (3)  ∃𝑅.𝐶⊑𝐷 (𝑐, 𝑑, 𝑟) = max ( 0, ( ‖ ‖ 𝑒 𝑣 (𝑐) − 𝑒 𝑣 (𝑟) − 𝑒 𝑣 (𝑑) ‖ ‖ − 𝑒 𝑟 (𝑐) − 𝑒 𝑟 (𝑑) − 𝛾 ) ) + | | | ‖ ‖ 𝑒 𝑣 (𝑐) ‖ ‖ − 1 | | | + | | | ‖ ‖ 𝑒 𝑣 (𝑑) ‖ ‖ − 1 | | | (4)</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.2.">Handling Bottom Concept (⊥):</head><p>Recall from the discussion in Section 3 that the bottom concept can appear only on the right hand side of the first three normal forms <ref type="bibr" target="#b7">[8]</ref>. We now present the loss functions for each of the three special cases. The resulting first normal form 𝐶 ⊑ ⊥ indicates that class 𝐶 is unsatisfiable. Thus, in the vector space, we represent this constraint by reducing the radius of class 𝐶 to zero. This is achieved by the following loss function.</p><formula xml:id="formula_5"> 𝐶⊑⊥ (𝑐) = 𝑒 𝑟 (𝑐)<label>(5)</label></formula><p>Next, the second normal form with the bottom concept is 𝐶 ⊓ 𝐷 ⊑ ⊥ indicating that 𝐶 and 𝐷 are disjoint. In the vector space, this indicates that the 𝑛-balls of classes 𝐶 and 𝐷 are nonoverlapping. This is captured by the following loss function.</p><formula xml:id="formula_6"> 𝐶⊓𝐷⊑⊥ (𝑐, 𝑑) = max ( 0, ( 𝑒 𝑟 (𝑐) + 𝑒 𝑟 (𝑑) − ‖ ‖ 𝑒 𝑣 (𝑐) − 𝑒 𝑣 (𝑑) ‖ ‖ + 𝛾 ) ) + | | | ‖ ‖ 𝑒 𝑣 (𝑐) ‖ ‖ − 1 | | | + | | | ‖ ‖ 𝑒 𝑣 (𝑑) ‖ ‖ − 1 | | |<label>(6)</label></formula><p>Finally, the third normal form ∃𝑅.𝐶 ⊑ ⊥ indicates that in the vector space translating 𝐶 by 𝑅 results in an unsatisfiable class. We already require the radius of unsatisfiable classes to be zero (Equation <ref type="formula" target="#formula_5">5</ref>) and since translation does not change the radius of the original class, we have the following loss function.</p><formula xml:id="formula_7"> ∃𝑅.𝐶⊑⊥ (𝑐, 𝑟) = 𝑒 𝑟 (𝑐)<label>(7)</label></formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.3.">Loss Functions for Role Inclusions and Role Chains:</head><p>The role vectors in our proposed framework serve the purpose of translating one class to another class. The constraints considered until now have imposed restrictions on the role vectors based on their relations with the 𝑛-balls of the concerned classes. We now present two loss functions to capture the constraints imposed by role inclusions and role chains in the ontology. The role inclusion of 𝑅 ⊑ 𝑆 implies that the vectors 𝑒 𝑣 (𝑅) and 𝑒 𝑣 (𝑆) in the vector space should be nearby because any translation produced by 𝑅 should also be producible by 𝑆 plus both the vectors should be in the same direction. This intuition is captured by the following loss function represented by Equation <ref type="formula" target="#formula_8">8</ref>. Herein, the first term is indicative of the distance that ensures the vectors 𝑒 𝑣 (𝑅) and 𝑒 𝑣 (𝑆) lie in near vicinity of each other. The second term captures the directional aspect of roles in vector space such that they tend to be in same direction.</p><formula xml:id="formula_8"> 𝑅⊑𝑆 (𝑟, 𝑠) = max ( 0, ‖ ‖ 𝑒 𝑣 (𝑠) − 𝑒 𝑣 (𝑟) ‖ ‖ − 𝛾 ) + | | | | | 1 − 𝑒 𝑣 (𝑟).𝑒 𝑣 (𝑠) ‖ ‖ 𝑒 𝑣 (𝑟) ‖ ‖ ‖ ‖ 𝑒 𝑣 (𝑠) ‖ ‖ | | | | | + | | | ‖ ‖ 𝑒 𝑣 (𝑟) ‖ ‖ − 1 | | | + | | | ‖ ‖ 𝑒 𝑣 (𝑠) ‖ ‖ − 1 | | |<label>(8)</label></formula><p>Next, we consider the hierarchy defined by the role chain 𝑅 1 •𝑅 2 ⊑ 𝑆. In the vector space, this implies that if class 𝐶 can be translated to class 𝐸 by successive application of 𝑅 1 and 𝑅 2 , it can also be translated to 𝐸 directly by the vector for role 𝑆 while preserving the direction of role vectors. The following loss function captures this behavior represented by Equation <ref type="formula" target="#formula_9">9</ref>.</p><formula xml:id="formula_9"> 𝑅 1 •𝑅 2 ⊑𝑆 (𝑟 1 , 𝑟 2 , 𝑠) = max ( 0, ‖ ‖ 𝑒 𝑣 (𝑠) − 𝑒 𝑣 (𝑟 1 ) − 𝑒 𝑣 (𝑟 2 ) ‖ ‖ − 𝛾 ) + | | | | | 1 − (𝑒 𝑣 (𝑟 1 ) + 𝑒 𝑣 (𝑟 2 )).𝑒 𝑣 (𝑠) ‖ ‖ (𝑒 𝑣 (𝑟 1 ) + 𝑒 𝑣 (𝑟 2 )) ‖ ‖ ‖ ‖ 𝑒 𝑣 (𝑠) ‖ ‖ | | | | | + | | | ‖ ‖ 𝑒 𝑣 (𝑟 1 ) ‖ ‖ − 1 | | | + | | | ‖ ‖ 𝑒 𝑣 (𝑟 2 ) ‖ ‖ − 1 | | | + | | | ‖ ‖ 𝑒 𝑣 (𝑠) ‖ ‖ − 1 | | |<label>(9)</label></formula><p>Often, negative sampling is employed during the training phase to learn better embeddings as negative samples can be easily generated to enhance the training data available. In order to incorporate negative samples in the training phase, the following loss function is employed.</p><formula xml:id="formula_10">𝑙𝑜𝑠𝑠 𝐶⋢∃𝑅.𝐷 (𝑐, 𝑑, 𝑟) = max ( 0, 𝑒 𝑟 (𝑐) + 𝑒 𝑟 (𝑑) − ‖ ‖ 𝑒 𝑣 (𝑐) + 𝑒 𝑣 (𝑟) − 𝑒 𝑣 (𝑑) ‖ ‖ + 𝛾 ) + | | | ‖ ‖ 𝑒 𝑣 (𝑐) ‖ ‖ − 1 | | | + | | | ‖ ‖ 𝑒 𝑣 (𝑑) ‖ ‖ − 1 | | |<label>(10)</label></formula><p>Thus, the total loss for learning the embedding function is the sum of all the loss functions given by Equations 1-10. Further, we also add the constraint that radius of the satisfiable classes are non-negative and penalize the total loss for learning negative radius for classes.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.3.">Training and Implementation</head><p>Given an  ++ ontology, we first normalize the ontology to generate the normal forms. These normal forms then constitute as a set of TBox statements wherein each axiom is treated as a positive sample. This normalization is performed using the OWL APIs and the APIs provided by the jCel reasoner which implements the normalization rules <ref type="bibr" target="#b17">[18]</ref>. We then introduce negative samples using the third normal form. We randomly generate corrupted axioms following 𝐶 ⊑ ∃𝑅.𝐷, by replacing 𝐶 or 𝐷 with 𝐶 ′ or 𝐷 ′ such that neither 𝐶 ′ ⊑ ∃𝑅.𝐷 nor 𝐶 ⊑ ∃𝑅.𝐷 ′ are asserted axioms in the ontology. Therefore, based on the facts the training process learns ontology embedding such that the facts hold true.</p><p>The code for training of embeddings and optimization is implemented using Python and Tensorflow library, and Adam optimizer <ref type="bibr" target="#b18">[19]</ref> is used for updating the embeddings. We start the learning process by initializing the embedding vectors for classes and relations by random values. We process the training samples in mini-batches for each of the losses defined for the normal forms along with the losses for roles, and update the embeddings depending upon the total loss i.e. the sum of all the loss functions. The update process is carried till saturation or a fixed number of epochs.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Experiments and Results</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Datasets</head><p>We use following four commonly used and publicly available ontologies of varying size and different characteristics. <ref type="bibr" target="#b19">[20]</ref> is one of the most comprehensive ontology of clinical terms with more than 989186 TBox statements involving 307712 classes and 60 relations.  <ref type="bibr" target="#b21">[22]</ref> unifies the representation of gene across all species. It consists of 130094 TBox statements, 45907 classes and 16 relations. 4. GALEN <ref type="bibr" target="#b22">[23]</ref> also represents clinical information. It consists of 84537 TBox statements with 24353 classes and 1010 relations.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">SNOMED CT</head><p>Table <ref type="table" target="#tab_1">1</ref> presents the four ontologies in the increasing order of their size (number of axioms) and highlights the differences between them in terms of the types of axioms. For instance, we note that SNOMED has 0 disjoint axioms and only 1 role chain axiom, despite being the largest amongst the four ontologies. On the other hand, GALEN, being one of the smaller ontologies, has the highest number of role inclusion (958) and role chain (58) axioms. Also, observe that ANATOMY is the only ontology considered that has the representation of all the EL constructs considered in this work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Baselines</head><p>We consider the following commonly used knowledge graph embeddings for comparison with our proposed EmEL ++ embeddings. <ref type="bibr" target="#b0">[1]</ref>, one of the most frequently used embedding model for knowledge graph, introduced the idea of translation based embeddings where the relations between entities is interpreted as a translation operation between the entities. 2. TransH <ref type="bibr" target="#b1">[2]</ref>, is an extension of TransE that better handles reflexive, one-to-many, manyto-one, and many-to-many relations. TransH considers relations as hyperplanes in the embedding space. The translation operation is then performed over the projections of entities on the hyperplane. 3. DistMult <ref type="bibr" target="#b2">[3]</ref>, a matrix factorization based embedding model, has been found empirically to perform well at compositional reasoning tasks. 4. EL Embeddings (ElEm) <ref type="bibr" target="#b6">[7]</ref> is one of the first embedding models for the  ++ description logic based model. Our proposed model is also an extension of ElEm embeddings and enhances ElEm by introducing additional constraints for a more comprehensive coverage of  ++ description logic. We use the pykeen framework <ref type="bibr" target="#b23">[24]</ref> for implementations of TransE, TransH, and DistMult embedding models. For ElEm embeddings, we used the source code provided by the authors<ref type="foot" target="#foot_0">1</ref> . Our implementation of EmEL ++ is publicly available at https://github.com/kracr/EmELpp.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">TransE</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Experimental Protocol</head><p>For learning the embeddings by different models, we first normalize the ontologies as described in Section 3. Next, we remove 30% of the subclass relation pairs from the normalized ontology to be used for validation (20%) and testing (10%). The remaining ontology with 70% sub-class relation pairs is used as the training set for learning the embedding functions. Further, we take an inferences set that consists of the inferences drawn on the training set using a standard Elk reasoner to evaluate the performance of learned embeddings. We perform hyper-parameter tuning using the 20% validation set and report the performance of fine-tuned models on the test set. The hyper-parameters to tune for all the models are the dimensions of the embedding vectors, and the margin parameter 𝛾 . We consider 𝑛 = {50, 100, 200} and 𝛾 = {−0.1, 0, 0.1} yielding nine different settings. The best performing hyper-parameters for each of the models are reported in Table <ref type="table" target="#tab_2">2</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4.">Reasoning Performance of EmEL ++</head><p>We chose subsumption as the main task to evaluate the effectiveness of the proposed EmEL ++ embeddings. Note that once we have embedded the ontologies in a vector space, we have to reduce all the tasks we want to accomplish to operations that can be performed in an 𝑛dimensional space. We reduce the task of subsumption in the embedding vector space as a distance-based operation. Given a test instance of the form 𝐶 ⊑ 𝐷, we take 𝐷 as our source class and rank all the other classes in the ontology in increasing order of their distance from 𝐷 in the vector space. We then compare the effectiveness of different embedding models based on the rank at which 𝐶 is present in the ranked list. An embedding model that successfully captures the subclass relation between the two classes should be able to assign vector representations to the two classes that are very close to each other, hence, producing a lower rank for 𝐶.</p><p>Table <ref type="table">3</ref> summarizes the performance of embedding models for test and inferences set. We report and evaluate the performance using six metrics. Hits at ranks 1, 10 and 100 report the fraction of test cases for which the expected class was found within top 1, 10 and 100 ranks, respectively. A median rank of 𝑚 means that for 50% of the test cases, the correct answer was</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Geometric Representation of classes and relation</figDesc><graphic coords="4,147.63,84.19,299.99,102.78" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head></head><label></label><figDesc>𝐶⊓𝐷⊑𝐸 (𝑐, 𝑑, 𝑒) = max ( 0, ( ‖ ‖ 𝑒 𝑣 (𝑐) − 𝑒 𝑣 (𝑑) ‖ ‖ − 𝑒 𝑟 (𝑐) − 𝑒 𝑟 (𝑑) − 𝛾 ) ) + max ( 0, ( ‖ ‖ 𝑒 𝑣 (𝑐) − 𝑒 𝑣 (𝑒) ‖ ‖ − 𝑒 𝑟 (𝑐) − 𝛾 ) ) + max ( 0, ( ||𝑒 𝑣 (𝑑) − 𝑒 𝑣 (𝑒)|| − 𝑒 𝑟 (𝑑) − 𝛾 ) ) + max ( 0, ( min ( 𝑒 𝑟 (𝑐), 𝑒 𝑟 (𝑑) ) − 𝑒 𝑟 (𝑒) − 𝛾 ) )</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1</head><label>1</label><figDesc>Different ontologies used in this work and count of different types of axioms. NF𝑖 represents the 𝑖 𝑡ℎ normal form as described in Section 3.</figDesc><table><row><cell>Ontology</cell><cell cols="2">SNOMED ANATOMY</cell><cell cols="2">GO GALLEN</cell></row><row><cell>Disjoint</cell><cell>0</cell><cell>184</cell><cell>30</cell><cell>0</cell></row><row><cell>Role Inclusion</cell><cell>11</cell><cell>89</cell><cell>3</cell><cell>958</cell></row><row><cell>Role Chain</cell><cell>1</cell><cell>31</cell><cell>6</cell><cell>58</cell></row><row><cell>NF1</cell><cell>446628</cell><cell cols="2">122142 85480</cell><cell>28890</cell></row><row><cell>NF2</cell><cell>27779</cell><cell cols="2">2121 12131</cell><cell>13595</cell></row><row><cell>NF3</cell><cell>482330</cell><cell cols="2">152289 20324</cell><cell>28118</cell></row><row><cell>NF4</cell><cell>32449</cell><cell cols="2">2143 12129</cell><cell>13597</cell></row><row><cell cols="5">2. Anatomy [21] is an ontology captures linkages of different phenotypes to genes. It</cell></row><row><cell cols="5">consists of 278883 TBox statements involving 106495 classes and 218 relations.</cell></row><row><cell>3. Gene Ontology(GO)</cell><cell></cell><cell></cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2</head><label>2</label><figDesc>Best performing Hyper-parameters for each model. 𝑛 indicates the dimension of embedding vectors and 𝛾 is the margin loss parameter.</figDesc><table><row><cell></cell><cell cols="2">EmEL ++</cell><cell>ElEm</cell><cell></cell><cell>TransE</cell><cell></cell><cell>TransH</cell><cell></cell><cell cols="2">DistMult</cell></row><row><cell></cell><cell>𝑛</cell><cell>𝛾</cell><cell>𝑛</cell><cell>𝛾</cell><cell>𝑛</cell><cell>𝛾</cell><cell>𝑛</cell><cell>𝛾</cell><cell>𝑛</cell><cell>𝛾</cell></row><row><cell>GALEN</cell><cell>50</cell><cell>0.0</cell><cell>50</cell><cell>0.0</cell><cell cols="2">100 -0.1</cell><cell cols="2">100 -0.1</cell><cell cols="2">100 -0.1</cell></row><row><cell cols="3">GO 100 -0.1</cell><cell cols="2">100 -0.1</cell><cell cols="2">100 -0.1</cell><cell cols="2">100 -0.1</cell><cell>100</cell><cell>0.1</cell></row><row><cell cols="3">ANATOMY 200 -0.1</cell><cell cols="2">200 -0.1</cell><cell cols="2">100 -0.1</cell><cell cols="2">100 -0.1</cell><cell cols="2">100 -0.1</cell></row><row><cell cols="3">SNOMED CT 100 -0.1</cell><cell cols="2">100 -0.1</cell><cell cols="2">50 -0.1</cell><cell cols="2">50 -0.1</cell><cell cols="2">50 -0.1</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://github.com/bio-ontology-research-group/el-embeddings</note>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>found below rank 𝑚. 90 𝑡ℎ percentile rank denotes the rank value below which the correct class was found for 90% of the test cases. The first observation that we make from Table <ref type="table">3</ref> is that ElEm and EmEL ++ embeddings perform better than the three commonly used KG embeddings (TransE, TransH, and DistMult). This observation highlights the inadequacy of traditional KG embeddings that do not consider the ontological constructs and rely only on the structural properties of the underlying graph. Both ElEm and EmEL ++ embeddings incorporate specific constraints and charactersitics of  ++ description logic, and hence, the embeddings produced by these models are better at retaining the properties of the underlying ontologies in the vector space. Next, we note from the two tables that there is no clear winner among ElEm and EmEL ++ . While EmEL ++ outperforms ElEm for the GALEN and GO ontologies, there is no clear winner for Anatomy and SNOMED ontologies as their performance varies. This observation is consistent with previous empirical studies comparing different link prediction methods that found that no single method outperforms across a variety of datasets <ref type="bibr" target="#b24">[25,</ref><ref type="bibr" target="#b25">26]</ref>. We speculate that this divergence in performance could be attributed to the different distributions of the types of axioms in the ontology (ref.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.5.">Preserving Ontology Characteristics in Vector Space</head><p>Next, we compare the ElEm and EmEL ++ in terms of retaining the underlying characteristics of ontology in vector space. Recall that both the models map the classes in an ontology to 𝑛-balls in the vector space. Further, the mapping is such that the 𝑛-ball of a super-class subsumes the 𝑛-balls of its sub-classes. Thus, for a test instance 𝐶 ⊑ 𝐷, we check that the 𝑛-ball of class 𝐶 lies inside the 𝑛-ball of class 𝐷 in the vector space. Note that since we have the centers and radii of the corresponding 𝑛-balls, this can be checked easily. Table <ref type="table">4</ref> presents the training, testing, and the inference accuracy obtained for the two embedding models for this task. We report accuracy values, i.e., the fraction of instances where the subsumption relation between the classes was maintained in the vector space. Note that this is a much stricter criterion for even if the subclass 𝑛-ball is slightly outside the 𝑛-ball of the superclass, it will be considered a failure. We observe from Table <ref type="table">4</ref> that EmEL ++ outperforms the ElEm embeddings for all the datasets and across all settings. This indicates that EmEL ++ embeddings are better at preserving the class relationships in the mapped vector space than ElEm embeddings.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Conclusions</head><p>We proposed EmEL ++ , an embedding model for  ++ ontologies. EmEL ++ builds upon and extends the previously proposed ElEm embeddings by focusing on role inclusions and role chains and offers a more complete coverage of  ++ . Experiments with four different ontologies showed that EmEL ++ outperforms traditional KB embeddings on the subsumption reasoning task. Further, when compared with ElEm embeddings, it is able to better preserve the underlying semantics of the ontologies in the vector space. We have also shown how to perform the subsumption reasoning task in a vector space, which is an 𝑂(𝑛) operation in the worst case. We believe this is an important capability and it offers exciting directions for future work.</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Translating embeddings for modeling multi-relational data</title>
		<author>
			<persName><forename type="first">A</forename><surname>Bordes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Usunier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Garcia-Duran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Weston</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Yakhnenko</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in neural information processing systems</title>
				<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="2787" to="2795" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Knowledge graph embedding by translating on hyperplanes</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Feng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Chen</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Twenty-Eighth AAAI conference on artificial intelligence</title>
				<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<author>
			<persName><forename type="first">B</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>-T. Yih</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Deng</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1412.6575</idno>
		<title level="m">Embedding entities and relations for learning and inference in knowledge bases</title>
				<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">A three-way model for collective learning on multi-relational data</title>
		<author>
			<persName><forename type="first">M</forename><surname>Nickel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Tresp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H.-P</forename><surname>Kriegel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Icml</title>
		<imprint>
			<biblScope unit="volume">11</biblScope>
			<biblScope unit="page" from="809" to="816" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Complex embeddings for simple link prediction</title>
		<author>
			<persName><forename type="first">T</forename><surname>Trouillon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Welbl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Riedel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">É</forename><surname>Gaussier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Bouchard</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Machine Learning</title>
				<imprint>
			<publisher>ICML</publisher>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Learning entity and relation embeddings for knowledge graph completion</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Twenty-ninth AAAI conference on artificial intelligence</title>
				<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">EL embeddings: geometric construction of models for the description logic el++</title>
		<author>
			<persName><forename type="first">M</forename><surname>Kulmanov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Liu-Wei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Yan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Hoehndorf</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 28th International Joint Conference on Artificial Intelligence</title>
				<meeting>the 28th International Joint Conference on Artificial Intelligence</meeting>
		<imprint>
			<publisher>AAAI Press</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="6103" to="6109" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Pushing the EL Envelope</title>
		<author>
			<persName><forename type="first">F</forename><surname>Baader</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Brandt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Lutz</surname></persName>
		</author>
		<idno>LTCS-05-01</idno>
		<ptr target="http://lat.inf.tu-dresden.de/research/reports.html" />
	</analytic>
	<monogr>
		<title level="m">Chair for Automata Theory</title>
				<imprint>
			<date type="published" when="2005">2005</date>
		</imprint>
		<respStmt>
			<orgName>Institute for Theoretical Computer Science, Dresden University of Technology, Germany</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">LTCS-Report</note>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Bernoulli embeddings for graphs</title>
		<author>
			<persName><forename type="first">V</forename><surname>Misra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Bhatia</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Thirty-Second AAAI Conference on Artificial Intelligence</title>
				<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">node2vec: Scalable feature learning for networks</title>
		<author>
			<persName><forename type="first">A</forename><surname>Grover</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Leskovec</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 22nd ACM SIGKDD international conference on Knowledge discovery and data mining</title>
				<meeting>the 22nd ACM SIGKDD international conference on Knowledge discovery and data mining</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="855" to="864" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Quantum Embedding of Knowledge for Reasoning</title>
		<author>
			<persName><forename type="first">D</forename><surname>Garg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Ikbal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">K</forename><surname>Srivastava</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Vishwakarma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Karanam</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">V</forename><surname>Subramaniam</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in Neural Information Processing Systems</title>
				<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="5595" to="5605" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Onto2vec: Joint vector-based representation of biological entities and their ontology-based annotations</title>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">Z</forename><surname>Smaili</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Gao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Hoehndorf</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Bioinformatics</title>
		<imprint>
			<biblScope unit="volume">34</biblScope>
			<biblScope unit="page" from="52" to="60" />
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Cone semantics for logics with negation</title>
		<author>
			<persName><forename type="first">Ö</forename><surname>Özçep</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Leemhuis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Wolter</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI</title>
				<meeting>the Twenty-Ninth International Joint Conference on Artificial Intelligence, IJCAI</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="1820" to="1826" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">Completion reasoning emulation for the description logic el+</title>
		<author>
			<persName><forename type="first">A</forename><surname>Eberhart</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ebrahimi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Zhou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Shimizu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Hitzler</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1912.05063</idno>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Pushing the EL envelope</title>
		<author>
			<persName><forename type="first">F</forename><surname>Baader</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Brandt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Lutz</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IJCAI</title>
		<imprint>
			<biblScope unit="volume">5</biblScope>
			<biblScope unit="page" from="364" to="369" />
			<date type="published" when="2005">2005</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Distributed representations of words and phrases and their compositionality</title>
		<author>
			<persName><forename type="first">T</forename><surname>Mikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sutskever</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">S</forename><surname>Corrado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Dean</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Advances in neural information processing systems</title>
				<imprint>
			<date type="published" when="2013">2013</date>
			<biblScope unit="page" from="3111" to="3119" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<analytic>
		<title level="a" type="main">Rdf2vec: Rdf graph embeddings for data mining</title>
		<author>
			<persName><forename type="first">P</forename><surname>Ristoski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Paulheim</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Semantic Web Conference</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="498" to="514" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<title level="m" type="main">jcel: A Modular Rule-based Reasoner</title>
		<author>
			<persName><forename type="first">J</forename><surname>Mendez</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2012">2012</date>
			<publisher>ORE</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m" type="main">Adam: a method for stochastic optimization</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">P</forename><surname>Kingma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Ba</surname></persName>
		</author>
		<idno>CoRR abs/1412.6980</idno>
		<imprint>
			<date type="published" when="2014">2014. 2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">SNOMED-CT: The advanced terminology and coding system for ehealth</title>
		<author>
			<persName><forename type="first">K</forename><surname>Donnelly</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Studies in health technology and informatics</title>
		<imprint>
			<biblScope unit="volume">121</biblScope>
			<biblScope unit="page">279</biblScope>
			<date type="published" when="2006">2006</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Uberon, an integrative multispecies anatomy ontology</title>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">J</forename><surname>Mungall</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Torniai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">V</forename><surname>Gkoutos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">E</forename><surname>Lewis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Haendel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Genome biology</title>
		<imprint>
			<biblScope unit="volume">13</biblScope>
			<biblScope unit="page">R5</biblScope>
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">The Gene Ontology (GO) database and informatics resource</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">O</forename><surname>Consortium</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Nucleic acids research</title>
		<imprint>
			<biblScope unit="volume">32</biblScope>
			<biblScope unit="page" from="D258" to="D261" />
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Rector</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Rogers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Pole</surname></persName>
		</author>
		<title level="m">The GALEN high level ontology</title>
				<imprint>
			<date type="published" when="1996">1996</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">The KEEN Universe</title>
		<author>
			<persName><forename type="first">M</forename><surname>Ali</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Jabeen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">T</forename><surname>Hoyt</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lehmann</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Semantic Web Conference</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="3" to="18" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">The link-prediction problem for social networks</title>
		<author>
			<persName><forename type="first">D</forename><surname>Liben-Nowell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kleinberg</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of the American society for information science and technology</title>
		<imprint>
			<biblScope unit="volume">58</biblScope>
			<biblScope unit="page" from="1019" to="1031" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Link prediction in complex networks: A survey</title>
		<author>
			<persName><forename type="first">L</forename><surname>Lü</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Zhou</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Physica A: statistical mechanics and its applications</title>
		<imprint>
			<biblScope unit="volume">390</biblScope>
			<biblScope unit="page" from="1150" to="1170" />
			<date type="published" when="2011">2011</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
