<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">A System for Reasoning-based Link Prediction in Large Knowledge Graphs</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Hong</forename><surname>Wu</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">College of Intelligence and Computing</orgName>
								<orgName type="institution">Tianjin University</orgName>
								<address>
									<country key="CN">China</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Zhe</forename><surname>Wang</surname></persName>
							<affiliation key="aff1">
								<orgName type="department">School of Information and Communication Technology</orgName>
								<orgName type="institution">Griffith University</orgName>
								<address>
									<country>Australian</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Xiaowang</forename><surname>Zhang</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">College of Intelligence and Computing</orgName>
								<orgName type="institution">Tianjin University</orgName>
								<address>
									<country key="CN">China</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Pouya</forename><forename type="middle">Ghiasnezhad</forename><surname>Omran</surname></persName>
							<affiliation key="aff2">
								<orgName type="department">Research School of Computer Science</orgName>
								<orgName type="institution">Australian National University</orgName>
								<address>
									<country key="AU">Australia</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Zhiyong</forename><surname>Feng</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">College of Intelligence and Computing</orgName>
								<orgName type="institution">Tianjin University</orgName>
								<address>
									<country key="CN">China</country>
								</address>
							</affiliation>
						</author>
						<author role="corresp">
							<persName><forename type="first">Kewen</forename><surname>Wang</surname></persName>
							<email>k.wang@griffith.edu.au</email>
							<affiliation key="aff1">
								<orgName type="department">School of Information and Communication Technology</orgName>
								<orgName type="institution">Griffith University</orgName>
								<address>
									<country>Australian</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">A System for Reasoning-based Link Prediction in Large Knowledge Graphs</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">268BF7EFB68F82F19D9944FBF819CA95</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T14:06+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>This poster paper presents an efficient method R-Linker for link prediction in large knowledge graphs, based on rule learning. The scalability and efficiency is achieved by a combination of several optimisation techniques. Experimental results show that R-Linker is able to handle KGs with over 10 million of entities and more efficient than existing state-of-the-art methods including RLvLR and AMIE+ in rule learning stage for link prediction.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Introduction</head><p>Knowledge graphs (KGs), a new generation of knowledge bases, have received significant attention in semantic technologies. As a KG is usually very large (of size over 10 million entities), it is infeasible for manual construction. Also, KGs are usually incomplete. Thus, it is useful and challenging to automatically construct and enrich KGs. Link prediction is one of important tasks for automated construction of KGs. Given an entity e and a (binary) relation R, the problem of link prediction is to find an entity e such that the triple (e, R, e ) (or equivalently, the fact R(e, e )) is in the KG. A large number of methods for link prediction have bee proposed in the literature, including Neural LP, Node+LinkFeat and DISMULT <ref type="bibr" target="#b0">[1]</ref>. However, most of these methods work only for relatively small KGs like WN18 and FB15K.</p><p>AMIE+ <ref type="bibr" target="#b3">[4]</ref> and RLvLR <ref type="bibr" target="#b5">[6]</ref> are among more recent methods that are able to predict links for larger KGs of size over 10 millions, and thus these methods are much more scalable than other rule learners such as <ref type="bibr" target="#b2">[3,</ref><ref type="bibr" target="#b4">5]</ref>. As these two methods are essentially rule learners, they can address the link prediction in a more general form. For convenience, we refer this link prediction as Reasoning-based Link Prediction or just R-link prediction. Specifically, given a relation R, we want to find a pair (or pairs) e and e of entities such that triple (e, R, e ) is in the KG. In particular, RLvLR demonstrates that the technique of embedding in representation learning is promising for handling R-link prediction problem in large KGs. In order to extract information on the nationality of persons, one can learn a rule like BornIn(x, y) ∧ Country(y, z) → Nationality(x, z). Rules are explicit knowledge (compared to a neural network) and can provide human understandable explanations to learning results (e.g., link prediction) based on them. Thus, it is useful and important to extract rules for KGs automatically.</p><p>In this poster paper, we further push the envelope by developing a more efficient method R-Linker for R-link prediction in large KGs. The scalability and efficiency of R-Linker is achieved by a combination of several optimisation techniques. First, we use an adapted embedding for rule learning; Second, we introduce a new strategy of sampling called Hierarchical Sampling; Moreover, we develop new techniques of rule search and rule evaluation. As a result, we have implemented a new system R-Linker for link prediction with large KGs. Our experiments show that R-Linker is able to handle KGs of size over 10M and more efficient than other methods including RLvLR and AMIE+ in rule learning stage for link prediction. R-Linker is available at https: //www.dropbox.com/sh/c8ent25u3qp4vp1/AABc6Jl3zTRtOkdTwHaoDBDUa?dl=0</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">A Rule-based Model</head><p>Unlike other statistical relational models, we adopt rule-based models for link prediction, with the obvious advantage that the learned models (as sets of logical rules) are explainable and reusable. In what follows, we describe how we construct such models.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1">Embedding-based Rule Selection</head><p>Inspired by <ref type="bibr" target="#b5">[6]</ref>, we learn such rule-based models via predicate embeddings; yet unlike <ref type="bibr" target="#b5">[6]</ref> using matrix embeddings, we adopt TransE vector embeddings which can significantly improve learning efficiency. As we demonstrate in the experiments, adopt a simpler form of embeddings does not compromise the learning accuracy. In <ref type="bibr" target="#b1">[2]</ref>, vector embeddings r and e are constructed for each relation R and each entity e in the KG. When a fact R(e, e ) exists in the KG, the embeddings satisfy e+r ≈ e . We extend it to an embedding characterisation for closed-path rules, that is, first-order Horn rules of the form R 1 (x, z 1 ) ∧ R 2 (z 1 , z 2 ) ∧ . . . ∧ R n (z n−1 , y) → R(x, y) with x, y, z 1 , . . . , z n−1 being variables. There are two aspects we hope to capture: (1) the composition of relations R 1 , . . . R n associates entities (in place of x and y) similarly as relation R does; and (2) the co-occurrence of arguments in the positions of x, y, z 1 , . . . , z n−1 .</p><p>For (1), it requires for each pair of entities (e, e ), e+r 1 +• • •+r n −e ≈ e+r−e . We define a measure sim(r 1 + • • • + r n , r), where sim is the L2 norm of vector distances. For (2), we use the notion of argumentation embeddings from <ref type="bibr" target="#b5">[6]</ref>. More specifically, for each relation R, two vector embeddings r 1 and r 2 are computed by averaging the entity embeddings (as vectors) of all the entities occurring in the position of respectively, the subject and object arguments of R. Then, for x occurring as the subject arguments of both R 1 and R, y occurring as the object argument of both R n and R, and z i (1 ≤ i ≤ n − 1) occurring as the object argument of R i and subject argument of R i+1 , we have the following measure sim(r</p><formula xml:id="formula_0">1 1 , r 1 ) + sim(r 2 n , r 2 ) + sim(r 2 1 , r 1 2 ) + • • • + sim(r 2 n−1 , r 1 n ).</formula></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2">Hierarchical Data Sampling</head><p>A major challenge in the computation of embeddings is that existing methods cannot scale over large KGs, even for vector embeddings. Hence, we propose a new data sampling strategy, called hierarchical sampling, to reduce the sizes of input KGs by focusing on entities that are relevant to the link prediction task. Intuitively, for each link prediction task, the link (i.e., a relation) R is often given, and we sample entities (and facts) in the KG that are directly or indirectly related to R for embedding construction. Consider a KG K = (E, F ) with E being the set of all entities and F being the set of all facts (i.e., triples) in K. Our sampling method selects a (small) subset E ⊆ E that are relevant to R and focus on the facts F only about E (not mentioning other entities). Since each rule in our model forms a path, our sampling method also deploys a breath-first tree search. As shown in Figure <ref type="figure" target="#fig_0">1</ref> (a), the first sampled entities E 0 are those occurring in facts about R. Then, E 1 are those entities that occur in any facts (not necessarily about R) mentioning entities from E 0 . Similarly, E i+1 are those entities that occur in any facts mentioning entities from E i , for each i ≥ 1 till a prescribed depth. Figure <ref type="figure" target="#fig_0">1</ref> (b) shows how our sampling method preserves closed-path rules. For a rule of length 3, R 1 (x, z) ∧ R 2 (z, y) → R(x, y) and each supporting instance of the rule R 1 (e, e ) ∧ R 2 (e , e ) → R(e, e ), entities e and e will be sampled in E 0 and e in E 1 .</p><formula xml:id="formula_1">𝑹 𝑬 𝟏 𝑬 𝟑 𝑬 𝟐 𝑬 𝟏 𝑬 𝟎 𝑬 𝟎 R 1 R 1 𝑹 𝐄 𝟐 𝑬 𝟏 𝑬 𝟎 𝑬 𝟐 𝑹 2 𝑹 1 𝑹 2 𝑹 1 𝑬 𝟎 𝑬 𝟏 𝑬 𝟏 𝑹 1 𝑹 2 𝑹 1 𝑬 𝟎 𝑬 𝟏 𝑬 𝟎 𝑙𝑒𝑛 = 2 𝑙𝑒𝑛 = 3 𝑙𝑒𝑛 = 4 𝑹 𝑹</formula><p>One optimisation is loop elimination during the breath-first search, as shown in Figure <ref type="figure" target="#fig_0">1</ref> (a), if a repeated entity is found on a path (represented in light color), the path is no longer explored. This is to avoid redundant atoms in the rules, for example R 1 (x, y) ∧ R − 1 (y, x) ∧ R 1 (x, y) → R(x, y). Furthermore, by recording the path information during the search, it eliminates a large number of invalid compositions of relations and can effectively suggest candidate rules. Other optimisations include selecting a bounded number of neighbours for each entity, and pruning relations with low frequency.</p><p>The evaluation of candidate rules, through the computation of standard confidence and head coverage, is often expensive, and much research effort has be dedicated to optimise such computation. A key step is to compute the support degree, i.e., the number of entity pairs in KG that make both body and head of the rule true. From the above discussions, we can quickly narrow down our search to entities directly connected to those in E 0 , and since the relation in the head is known, we can first check whether a pair of entities satisfy the head. These optimisations prove to be quite effective.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Experiments</head><p>We compared our system with RLvLR, AMIE+ and Neural LP on rule learning and link prediction, on common benchmarks FB15K(-237), Wikidata, DBPedia 3.8, and YAGO2s.</p><p>For large KGs Wikidata, DBPedia 3.8, and YAGO2s, Table <ref type="table" target="#tab_0">1</ref> shows our system outperforms both RLvLR and AMIE+ in learning efficiency, as shown by the average numbers of rules (#R) and quality rules (#QR, standard confidence over 0.7) learned per hour. Table <ref type="table">2</ref> shows that compared to RLvLR and Neural LP, the model constructed by our system demonstrates better accuracy on link prediction. Table <ref type="table">2</ref> shows the comparison of our rule-based model against statistical models on FB15K-237. While our model has competitive performance on link prediction, its major advantage is that rule-based models are explainable and reusable. We plan to compare our method with some other approaches such as <ref type="bibr" target="#b6">[7]</ref>.</p><p>Table <ref type="table">2</ref>: Link prediction on large KGs.</p><p>Learner FB75K Wikidata MRR Hits@10 MRR Hits@10 R-Linker 0.37 59.0 0.33 39. </p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Fig. 1 :</head><label>1</label><figDesc>Fig. 1: (a) Breath-first search for sampling; (b) Sampling preserves closed-paths.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1 :</head><label>1</label><figDesc>Rule learning on large KGs.</figDesc><table><row><cell>KG</cell><cell cols="3">R-Linker #R #QR #R #QR #R #QR RLvLR AMIE+</cell></row><row><cell cols="2">DBpedia 13.38 3.67</cell><cell>11 2.37 1.97</cell><cell>0.11</cell></row><row><cell cols="4">Wikidata 37.38 18.52 23.56 10.62 &lt;0.09 &lt;0.03</cell></row><row><cell cols="4">YAGO2s 9.71 2.28 6.56 1.88 &lt;0.56 &lt;0.05</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 3 :</head><label>3</label><figDesc>Link prediction on FB15K-237.</figDesc><table><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>Learner</cell><cell cols="2">MRR Hits@10</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>DISTMULT</cell><cell>0.25</cell><cell>40.8</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell>3</cell><cell cols="2">Node+LinkFeat 0.23</cell><cell>34.7</cell></row><row><cell>RLvLR</cell><cell>0.34</cell><cell>43.4</cell><cell>0.29</cell><cell>38.9</cell><cell>Neural LP</cell><cell>0.24</cell><cell>36.1</cell></row><row><cell cols="2">Neural LP 0.13</cell><cell>25.7</cell><cell>-</cell><cell>-</cell><cell>RLvLR</cell><cell>0.24</cell><cell>39.3</cell></row><row><cell></cell><cell></cell><cell></cell><cell></cell><cell></cell><cell>R-Linker</cell><cell>0.24</cell><cell>38.1</cell></row></table></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">A Review of Relational Machine Learning for Knowledge Graph</title>
		<author>
			<persName><forename type="first">M</forename><surname>Nickel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Murphy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Tresp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Gabrilovich</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of IEEE</title>
				<meeting>IEEE</meeting>
		<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="volume">1041</biblScope>
			<biblScope unit="page" from="1" to="23" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Translating embeddings for modeling multi-relational data</title>
		<author>
			<persName><forename type="first">B</forename><surname>Antoine</surname></persName>
		</author>
		<author>
			<persName><forename type="first">U</forename><surname>Nicolas</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">D</forename><surname>Alberto</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Jason</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Oksana</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">NIPS</title>
		<imprint>
			<biblScope unit="volume">26</biblScope>
			<biblScope unit="page" from="2787" to="2795" />
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Scalekb: scalable learning and inference over large knowledge bases</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">Z</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Goldberg</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The VLDB Journal</title>
		<imprint>
			<biblScope unit="volume">25</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="893" to="918" />
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Fast rule mining in ontological knowledge bases with amie+</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">A</forename><surname>Galárraga</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Teflioudi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Hose</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">M</forename><surname>Suchanek</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The VLDB Journal</title>
		<imprint>
			<biblScope unit="volume">24</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page" from="707" to="730" />
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Rule learning from knowledge graphs guided by embedding models</title>
		<author>
			<persName><forename type="first">V</forename><forename type="middle">T</forename><surname>Ho</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Stepanova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">H</forename><surname>Gad-Elrab</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Kharlamov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Weikum</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. ISWC pp</title>
				<meeting>ISWC pp</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="72" to="90" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Scalable rule learning via learning representation</title>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">G</forename><surname>Omran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. AAAI</title>
				<meeting>AAAI</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="2149" to="2155" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Blrn: End-to-end learning of knowledge base representations with latent, relational, and numerical features</title>
		<author>
			<persName><forename type="first">A</forename><surname>García-Durán</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Niepert</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proc. UAI</title>
				<meeting>UAI</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="372" to="381" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
