<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Knowledge Graph Embeddings or Bias Graph Embeddings? A Study of Bias in Link Prediction Models</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Andrea</forename><surname>Rossi</surname></persName>
							<email>andrea.rossi3@uniroma3.it</email>
							<affiliation key="aff0">
								<orgName type="institution">Roma Tre University</orgName>
								<address>
									<settlement>Roma</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Donatella</forename><surname>Firmani</surname></persName>
							<email>donatella.firmani@uniroma1.it</email>
							<affiliation key="aff1">
								<orgName type="institution">Sapienza University</orgName>
								<address>
									<settlement>Roma</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Paolo</forename><surname>Merialdo</surname></persName>
							<email>paolo.merialdo@uniroma3.it</email>
							<affiliation key="aff0">
								<orgName type="institution">Roma Tre University</orgName>
								<address>
									<settlement>Roma</settlement>
									<country key="IT">Italy</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Knowledge Graph Embeddings or Bias Graph Embeddings? A Study of Bias in Link Prediction Models</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">62D0008E021DFFD3B0722A373F41C34A</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T14:03+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>Link Prediction, Knowledge Graph Embeddings</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Link Prediction aims at tackling Knowledge Graph incompleteness by inferring new facts based on the existing, already known ones. Nowadays most Link Prediction systems rely on Machine Learning and Deep Learning approaches; this results in inherent opaque models in which assessing the robustness to data biases is not trivial. We define 3 specific types of Sample Selection Bias and estimate their presence in the 5 best-established Link Prediction datasets. We then verify how these biases affect the behaviour of 9 systems representative for every major family of Link Prediction models. We find that these models do indeed learn and incorporate each of the presented biases, with a heavily negative effect on their behaviour. We thus advocate for the creation of novel more robust datasets and of more effective evaluation practices.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Knowledge Graphs (KGs) are structured repositories of information where nodes modeling real-world entities are linked by labeled directed edges; each label represents a semantic relation, therefore each edge linking a pair nodes represents a fact conveying that the corresponding two entities are connected via that relation.</p><p>KGs have recently achieved widespread popularity in a variety of contexts. Large open KGs, such as DBpedia <ref type="bibr" target="#b0">[1]</ref>, YAGO <ref type="bibr" target="#b1">[2]</ref> and Wikidata <ref type="bibr" target="#b2">[3]</ref>, are used on a daily basis for Semantic Web projects <ref type="bibr" target="#b3">[4]</ref> and question answering <ref type="bibr" target="#b4">[5]</ref>. Meanwhile, a growing number of companies rely on private KGs to support their services. Google and Microsoft use respectively the Google KG <ref type="bibr" target="#b5">[6]</ref> and Satori <ref type="bibr" target="#b6">[7]</ref> to enhance their search engines; Amazon <ref type="bibr" target="#b7">[8]</ref> and Ebay <ref type="bibr" target="#b8">[9]</ref> use product graphs to improve their recommendations; social networks like Facebook <ref type="bibr" target="#b9">[10]</ref> and LinkedIn <ref type="bibr" target="#b10">[11]</ref> use KGs for user profiling and advertisement.</p><p>It is therefore unsurprising that many KGs have achieved web-scale dimensions, featuring millions of entities and billions of facts. Nonetheless, it is well-known that even even the largest and richest KGs suffer from incompleteness, as they only hold a small portion of the real-world knowledge they should encompass <ref type="bibr" target="#b11">[12]</ref>.</p><p>Link Prediction (LP) tackles this issue by leveraging the already known facts in the KG to infer new ones. Nowadays the vast majority of LP models learn vectorized representations of entities and relations called embeddings using Machine Learning (ML) techniques; many of them rely on Deep Learning architectures, featuring sequences of neural layers interspersed by activation functions. These models have been shown to achieve state-of-the-art performances <ref type="bibr" target="#b12">[13,</ref><ref type="bibr" target="#b13">14]</ref>. In the last decade embedding-based LP has become a sparkling research area, with dozens of novel models being proposed every year (see the work by Wang et al. <ref type="bibr" target="#b14">[15]</ref> for a comprehensive survey). The pioneering model TransE <ref type="bibr" target="#b15">[16]</ref> has established the practice of evaluating these systems computing global metrics of their predictive performances over datasets obtained from real-world KGs.</p><p>LP datasets are usually obtained by extracting the most mentioned entities from real-world KGs. We find that this policy leads to various forms of imbalances and biases. For instance, the datasets sampled from Freebase <ref type="bibr" target="#b16">[17]</ref> tend to only feature people with nationality 𝑈 𝑆𝐴. Furthermore, the same biases observed in training are also present in validation and testing: this indirectly incentivizes models to incorporate the biases, as they can be instrumental to produce the correct predictions and boost the evaluation results. In short, our models are yielding the correct answers for the wrong reasons. For instance, in FB15k-237, which is considered the most reliable dataset <ref type="bibr" target="#b12">[13]</ref>, among the 1210 training facts and 151 test facts with relation film_budget_currency, 1164 and 146 facts respectively feature the same tail USA_dollar. We find that in those test facts, models always predict the correct tail on the first try when the answer is USA_dollar, whereas they never manage to guess if it is a different currency.</p><p>A few studies in the past have highlighted criticalities in LP benchmarks, mainly with regard to test leakage <ref type="bibr" target="#b17">[18,</ref><ref type="bibr" target="#b18">19]</ref> or of unnatural distributions and structures <ref type="bibr" target="#b19">[20,</ref><ref type="bibr" target="#b12">13]</ref>. Nonetheless, to the best of our knowledge the presence of straight-up biases had gone almost unnoticed so far. This motivates a systematic analysis of how they affect datasets and models, especially considering that KG embeddings have been shown to be just as vulnerable to biases as word embeddings <ref type="bibr" target="#b20">[21,</ref><ref type="bibr" target="#b21">22]</ref>. We report an accurate comparison with these other researches in Section 5.</p><p>We provide a formal definition of 3 types of data bias that can affect LP models. We focus on the 5 best-established LP datasets and estimate for each of them the number of test samples affected by each type of bias. We then conduct an extensive re-evaluation of 9 models representative for all the major families of LP systems, showing how removing the biased test facts affects their overall predictive performance. All the code and resources generated in our work are available at our public repository. <ref type="foot" target="#foot_0">1</ref>The paper is organized as follows. In Section 2 we overview how embedding-based models perform the LP task; in Section 3 we discuss the main types of data bias we analyze in our work; in Section 4 we report our experimental findings on how they affect LP models; in Section 5 we present works related to ours and in Section 6 we provide concluding remarks.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Link Prediction on Knowledge Graphs</head><p>We define any Knowledge Graph as 𝐾𝐺 = (ℰ, ℛ, 𝒢), where ℰ is a set of entities, ℛ is a set of relations, and 𝒢 ∈ ℰ × ℛ × ℰ is a set of facts connecting the entities via relations. Each fact can be formulated as a triple ⟨ℎ, 𝑟, 𝑡⟩, where ℎ is the head, 𝑟 is the relation, and 𝑡 is the tail.</p><p>Most LP models nowadays map entities and relations to vectorized representations called KG embeddings. These models usually define a scoring function 𝜑 to estimate the plausibility of facts based on the embeddings of their elements. Embeddings are initialized randomly; then, they are trained with ML methods to optimize the scores of the known facts. When the training is over, the learned embeddings should be able to generalize and yield good 𝜑 values for unseen true facts as well. Models may also feature deep architectures of neural layers, which can be used in 𝜑 to process the embeddings of elements of the facts to score. The weights of neural layers are trained jointly with the KG embeddings.</p><p>Given a trained model, a tail prediction ⟨ℎ, 𝑟, 𝑡⟩ is the process that finds 𝑡 to be the bestscoring entity to complete the triple ⟨ℎ, 𝑟, ?⟩: i.e., 𝑡 is the answer to the question «What is the most likely tail for head ℎ and relation 𝑟?»<ref type="foot" target="#foot_1">2</ref> :</p><formula xml:id="formula_0">𝑡 = argmax 𝑒∈ℰ 𝜑(ℎ, 𝑟, 𝑒).<label>(1)</label></formula><p>A formulation for head predictions can be defined analogously. LP research typically relies on datasets sampled from real-world KGs. Any dataset has its own sets ℰ, ℛ, and 𝒢, and 𝒢 is usually split into a training set 𝒢 𝑡𝑟𝑎𝑖𝑛 , a validation set 𝒢 𝑣𝑎𝑙𝑖𝑑 and a test set 𝒢 𝑡𝑒𝑠𝑡 . To evaluate the predictive performance of a model on a dataset, head and tail predictions are performed on each test fact in its 𝒢 𝑡𝑒𝑠𝑡 . For each prediction, the target entity featured in the test fact (i.e., the expected prediction) is ranked against all other entities in ℰ. Given any test fact ⟨ℎ, 𝑟, 𝑡⟩, the tail rank can be thus computed as:</p><formula xml:id="formula_1">𝑡𝑎𝑖𝑙𝑅𝑎𝑛𝑘(ℎ, 𝑟, 𝑡) = |{𝑒 ∈ ℰ|𝜑(ℎ, 𝑟, 𝑒) &gt;= 𝜑(ℎ, 𝑟, 𝑡)}| (2)</formula><p>Head ranks can be computed analogously. An ideal model, or an oracle, would obtain rank 1 in the head and tail predictions of all test facts. The set 𝑇 of all the head and tail ranks obtained in testing are then gathered into global metrics:</p><p>• Hits@K (H@K): the fraction of ranks in 𝑇 lesser or equal than a value 𝑘.</p><p>• Mean Reciprocal Rank (MRR): the average of the inverse values of all the ranks in 𝑇 .</p><p>Both 𝐻@𝐾 and 𝑀 𝑅𝑅 are always between 0 and 1; the higher their value, the better the result they convey. These metrics can be computed in two separate settings: in the raw setting correct answers that outrank the target one are still deemed wrong; in the filtered setting they are not considered mistakes and do not contribute to rank computation. Raw metrics can sometimes be misleading, so filtered metrics are generally preferred in literature <ref type="bibr" target="#b15">[16]</ref>; therefore, in this work we always use filtered metrics.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Forms of Data Bias</head><p>In this section, we define 3 main types of bias commonly found in LP datasets. All of them are forms of Sample Selection Bias <ref type="bibr" target="#b22">[23]</ref>, i.e., unwanted, unrealistic patterns in a dataset caused by imbalances in the processes and sources used to construct it. We provide definitions from the perspective of a tail prediction ⟨ℎ, 𝑟, 𝑡⟩; analogous definitions can be used for head predictions. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Data Bias in Link Prediction Benchmarks</head><p>We discuss in this section our experimental findings on the three forms of data bias defined in Section 3. We first describe how these biases affect LP datasets; we then analyze their consequences on the behaviour of LP models.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Data Biases in LP Datasets</head><p>We take into account the 5 most popular LP datasets in literature: FB15k, WN18, FB15k-237, WN18RR and YAGO3-10. FB15k and WN18 have been built by Bordes et al. <ref type="bibr" target="#b15">[16]</ref> by selecting the facts featuring the richest entities in Freebase <ref type="bibr" target="#b16">[17]</ref> and WordNet <ref type="bibr" target="#b23">[24]</ref>. Toutanova and Chen <ref type="bibr" target="#b17">[18]</ref> and Dettmers et al. <ref type="bibr" target="#b18">[19]</ref>  have also created the YAGO3-10 dataset selecting the facts featuring entities linked by at least 10 different relations in the YAGO3 KG <ref type="bibr" target="#b24">[25]</ref>.</p><p>In these datasets the training, validation and test sets are sampled from the same distributions, so any bias seen in training will also be featured in validation and testing. For each test prediction in each dataset we verify if it is prone to any type of bias applying the definitions in Section 3. Table <ref type="table" target="#tab_1">1</ref> reports the main statistics and the number of test predictions not affected by Type 1 Bias (w/o B1), Type 2 Bias (w/o B2), Type 3 Bias (w/o B3), and by any type of bias at all (w/o B⋆). Note that these numbers are not cumulative because the same prediction may be affected by multiple types of bias.</p><p>The most affected dataset is YAGO3-10, with 54.1% of its test facts being prone to bias, especially of Type 3. This confirms the finding of Akrami et al <ref type="bibr" target="#b12">[13]</ref> that the two most common relations in the dataset (affiliated_to and plays_for) are almost interchangeable. FB15k and FB15k-237 are noticeably affected too, with 26.7% and 9.8% of their test facts prone to bias respectively. The Type 3 Bias is not present at all in FB15k-237; this is not surprising, because this dataset was built by design with no inverse or equivalent relations. WN18 and WN18RR, finally, seem completely immune to any type of bias; this is possibly due the nature of WordNet, which is a lexical ontology rather than a KG.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Data Biases and LP models</head><p>In this section we study how data bias affects the behaviour of LP models. As mentioned in Section 4.1, LP benchmarks display the same biases across their training and test sets. Hence, in addition to being exposed to bias, unbeknownst to developers, LP models are actually encouraged to incorporate it, because such bias can help to identify the correct answers in testing. In the long run this may favour the architectures most vulnerable to bias over the most robust ones.</p><p>To assess the severity of this condition we remove from each dataset the test predictions prone to each type of bias, and measure how this affects the evaluation results of LP models relying on diverse architectures. We focus on H@1 and MRR, usually considered the most characterizing metric in LP. We stress that we do not modify the training sets of the datasets: we just filter away the bias-prone test predictions, in the quantities reported in Table <ref type="table" target="#tab_1">1</ref>.</p><p>We use as a starting point the evaluation results of 19 LP models publicly shared by Rossi et al. <ref type="bibr" target="#b13">[14]</ref>. Due to space constraints, we report our findings on 9 models represent- </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Key Takeaways</head><p>We find that the policies used to generate LP datasets have led to severe forms of selection bias; this, in turn, has made significant fractions of their test sets artificially easier to predict than the others. Our results prove that LP models are indeed sensitive to these forms of bias, as filtering away the affected test predictions heavily worsens their evaluation metrics. Both neural and rule-based LP models appear equally vulnerable to this phenomenon: this proves that the issue is not rooted in the technique used to learn the facts, but rather in the data sources themselves. Not all datasets are interested by this condition to the same extent. YAGO3-10 and FB15k-237 are the most affected, and show the heaviest drop in performance when removing the biased test facts. Wordnet-based datasets, on the other hand, do not display bias at all.</p><p>The bias problem, in itself, can probably be avoided by just skipping the test facts prone to bias during evaluation. However, suggesting to just adopt this practice would be naive from our part. Quite worryingly, even in bias-free datasets we observe that most correct predictions are just enabled by the presence of inverse and symmetric relations. In other words, in 𝑎𝑙𝑙 datasets, when removing the test predictions affected by either bias or inverse/symmetric relations, the predictive performance of models plummets.</p><p>This makes us wonder how many of the remaining test facts are actually predictable. If we just removed the bias-prone test predictions, we may mostly end up with test sets that not even humans, with the information available in training, can infer; if this was the case, the whole task would become pointless. We intend to conduct further studies in this regard, asking this question directly to human workers. If the outcome will prove that most non-biased test facts are indeed unpredictable, then it will be painfully necessary to replace the current datasets with novel ones extracted in more sensible ways, possibly keeping humans in the loop for the selection of training and test facts.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Related Works</head><p>The works most related to ours consist in analyses that highlight criticalities of the current LP benchmarking techniques. So far, compared to the wide body of literature proposing new LP models, a relatively small effort has been devoted to analyzing their evaluation practices.</p><p>Toutanova and Chen <ref type="bibr" target="#b17">[18]</ref> have been the first to notice the presence of test leakage in FB15k due to inverse relations. They have assessed the severity of the issue by proving that a simple model based on observable features achieves competitive performance on the dataset; they have proceeded to remove these relations from FB15k generating the more challenging FB15k-237.</p><p>A similar study has been carried out by Dettmers et al. <ref type="bibr" target="#b18">[19]</ref> on FB15k and WN18, showing that a trivial system based on inverse relations can achieve state-of-the-art results on both datasets; the authors have then generated WN18RR as a more challenging subsample of WN18.</p><p>Akrami et al. <ref type="bibr" target="#b12">[13]</ref> have carried out an extensive analysis quantifying the effect of various artificial patterns in LP datasets, such as inverse relations and Cartesian product relations. In addition to confirming the above mentioned observations on FB15k and WN18 they have found that LP performances are boosted in FB15k and FB15k-237 by the redundant structures of Cartesian product relations, and in YAGO3-10 by the presence of equivalent relations.</p><p>Nayyeri et al. <ref type="bibr" target="#b33">[34]</ref> refer to KGs in which facts also feature additional numerical weights to convey various meanings, e.g., the confidence of each fact. They acknowledge that the presence of bias in these values hinders the effectiveness of the learned embeddings, and propose a Weighted Triple Loss that, while taking advantage of these weights, is also robust to their biases.</p><p>Rossi and Matinata <ref type="bibr" target="#b19">[20]</ref> have shown that the distributions of entities in all LP datasets are wildly skewed: a few entities are featured in thousands of training facts, making them easier to learn and predict, whereas the others may only occur a handful of times. The rich and easily predictable entities are over-represented in testing, thus affecting fairness of such benchmarks.</p><p>The same concerns have been shared by Mohamed et al. <ref type="bibr" target="#b34">[35]</ref>; to overcome this issue, they have proposed novel stratified versions of the Hits@K and MRR metrics, called Strat-Hits@K and Strat-MRR respectively. These metrics should estimate of the predictive performance of models in a fairer way, unbiased by the popularity over-represented entities.</p><p>All these works share the same spirit of ours in their goal to identify the shortcomings of current LP evaluation approaches, in order to drive research towards more realistic and healthy practices. Our main difference lies in our definition and identification of bias structures that had so far gone unnoticed, as well as our systematic methodology of re-computing the predictive performances of a wide variety of models after removing the biased test facts.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Conclusion</head><p>We have reported an analysis on the presence of bias across the 5 best-established datasets for Link Prediction on KGs. We have defined 3 main types of Selection Sample Bias, and we have observed that they affect significant portions of the test predictions in 3 datasets out of 5. We have then analyzed how removing such bias-prone predictions alters the evaluation results of 9 models representing the main families of Link Prediction systems. The result is generally a significant drop in predictive performance. This proves that a large part of the correct predictions output by models on those datasets is indeed facilitated by the presence of bias. The very low values obtained on this de-biased test scenario suggests that many of the remaining test facts may not be predictable at all.</p><p>We thus call for the production of more effective and robust datasets for Link Prediction, and for the definition of more thorough evaluation methods that take into account their properties.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>Type 1 Bias. A tail prediction ⟨ℎ, 𝑟, 𝑡⟩ is prone to Type 1 Bias if the training facts mentioning 𝑟 tend to always feature 𝑡 as tail. For example, the tail prediction ⟨𝐵𝑎𝑟𝑎𝑐𝑘_𝑂𝑏𝑎𝑚𝑎, 𝑔𝑒𝑛𝑑𝑒𝑟, 𝑚𝑎𝑙𝑒⟩ is prone to this type of bias if the vast majority of gendered entities in the training set are males: this artificially favours the prediction of male genders. In practice, we verify if the fraction between the number of training facts featuring both 𝑟 and 𝑡 and the number of training facts featuring 𝑟 exceeds a threshold 𝜏 1 . In our experiments we set 𝜏 1 = 0.75. Type 2 Bias. A tail prediction ⟨ℎ, 𝑟, 𝑡⟩ in which 𝑟 is a one-to-many or a many-to-many relation is prone to Type 2 Bias if, whenever an entity 𝑒 is seen as head for relation 𝑟, fact ⟨𝑒, 𝑟, 𝑡⟩ also exists in 𝒢 𝑡𝑟𝑎𝑖𝑛 . Type 2 Bias affects relations that have a "default" correct answer. Differently from Type 1, facts mentioning 𝑟 may feature a variety of tails different from 𝑡; however, for each entity 𝑒 seen as head these facts, 𝑡 tends to always be among the correct tails too. This makes ⟨𝑒, 𝑟, 𝑡⟩ artificially easier to predict. For instance, the tail prediction ⟨𝐶𝑟𝑖𝑠𝑡𝑖𝑎𝑛𝑜_𝑅𝑜𝑛𝑎𝑙𝑑𝑜, 𝑙𝑎𝑛𝑔𝑢𝑎𝑔𝑒, 𝐸𝑛𝑔𝑙𝑖𝑠ℎ⟩ is prone to Type 2 Bias if most people, in addition to other languages, also speak English. In practice, we verify if the fraction of entities 𝑒 seen as heads for relation 𝑟 and that also display a fact ⟨𝑒, 𝑟, 𝑡⟩ exceeds a threshold 𝜏 2 . In our experiments we use 𝜏 2 = 0.5.</figDesc><table /><note>Type 3 Bias. A tail prediction ⟨ℎ, 𝑟, 𝑡⟩ is prone to Type 3 Bias if a relation 𝑠 exists such that: (i) whenever 𝑠 links two entities, 𝑟 links them as well; and (ii) the fact ⟨ℎ, 𝑠, 𝑡⟩ is present in the training set. For example, in the FB15k dataset the producer of a TV program is almost always its creator too; this may lead to assume that creating a program implies being its producer. In practice, to verify if 𝑠 and 𝑟 share this correlation we check if the fraction of 𝑠 mentions in which 𝑠 also co-occurs with 𝑟 is greater than a threshold 𝜏 3 . In our experiments we set 𝜏 3 = 0.5.</note></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1</head><label>1</label><figDesc>have observed that such datasets suffer from test leakage due to the pervasive presence of inverse and equivalent relations; they have filtered away such relations to create the more challenging subsamples FB15k-237 and WN18RR. Finally, Dettmers et al.<ref type="bibr" target="#b18">[19]</ref> Dataset statistics and numbers of test predictions prone to bias.</figDesc><table><row><cell></cell><cell cols="2">Entities Relations</cell><cell>Train</cell><cell>Facts Valid</cell><cell>Test</cell><cell>All</cell><cell cols="4">Test Predictions w/o B1 w/o B2 w/o B3 w/o B✻</cell></row><row><cell>FB15k</cell><cell>14.951</cell><cell cols="2">1.345 483.142</cell><cell>50.000</cell><cell>59.071</cell><cell>118.142</cell><cell>115.165 (-2.5%)</cell><cell>108.705 (-8.0%)</cell><cell>101.406 (-14.2%)</cell><cell>93.005 (-21.3)</cell></row><row><cell>FB15k-237</cell><cell>14.541</cell><cell cols="2">237 272.115</cell><cell>17.535</cell><cell>20.466</cell><cell>40.932</cell><cell>39.145 (-4.4%)</cell><cell>36.482 (-5.5%)</cell><cell>40.932 (-0.0%)</cell><cell>36.912 (-9.8%)</cell></row><row><cell>WN18</cell><cell>40.943</cell><cell cols="2">18 141.442</cell><cell>5.000</cell><cell>5.000</cell><cell>10.000</cell><cell>10.000 (-0.0%)</cell><cell>10.000 (-0.0%)</cell><cell>10.000 (-0.0%)</cell><cell>10.000 (-0.0%)</cell></row><row><cell>WN18RR</cell><cell>40.943</cell><cell>11</cell><cell>86.835</cell><cell>3.034</cell><cell>3.134</cell><cell>6.268</cell><cell>6.268 (-0.0%)</cell><cell>6.268 (-0.0%)</cell><cell>6.268 (-0.0%)</cell><cell>6.268 (-0.0%)</cell></row><row><cell cols="2">YAGO3-10 123.155</cell><cell cols="2">37 1.078.868</cell><cell>5.000</cell><cell>5.000</cell><cell>10.000</cell><cell>9.725 (-2.8%)</cell><cell>9.992 (-0.1%)</cell><cell>4.876 (-51.2%)</cell><cell>4.593 (-54.1%)</cell></row></table></figure>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="1" xml:id="foot_0">https://github.com/merialdo/research.lpbias</note>
			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" n="2" xml:id="foot_1">In this formulation we assume higher 𝜑 scores convey better plausibility; analogous formulations are defined for models where higher scores convey worse plausibility.</note>
		</body>
		<back>
			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><p>ing all the families in their work: ComplEx <ref type="bibr" target="#b25">[26]</ref> and TuckER <ref type="bibr" target="#b26">[27]</ref> for the Matrix Decomposition models; TransE <ref type="bibr" target="#b15">[16]</ref>, CrossE <ref type="bibr" target="#b27">[28]</ref> and HAKE <ref type="bibr" target="#b28">[29]</ref> for the Geometric models; InteractE <ref type="bibr" target="#b29">[30]</ref>, RSN <ref type="bibr" target="#b30">[31]</ref> and CapsE <ref type="bibr" target="#b31">[32]</ref> for the Deep Learning models; and the rule-based AnyBURL <ref type="bibr" target="#b32">[33]</ref> as a baseline. Complete results for all the 19 models are available in our online repository.</p><p>Table <ref type="table">2</ref> shows the results for datasets FB15k, FB15k-237 and YAGO3-10. If models were immune to biases, removing bias-prone test predictions should not heavily affect their metrics. On the contrary, we observe impressive performance drops across all models. YAGO3-10 results display the largest worsening, with models losing 50% to 70% of their H@1 and MRR. In FB15k- </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Table 3</head><p>Variations in the evaluation results on WN18 and WN18RR after removing the test predictions prone to bias or influenced by inverse and symmetric relations.</p><p>237, often considered the most reliable dataset, their decrease is still between 20% and 35%. In FB15k they lose around 5% of the original metrics, but this apparent robustness is likely due to the presence of inverse relations facilitating predictions.</p><p>Table <ref type="table">3</ref> reports the results for WN18 and WN18RR. As already described in Section 4.1, in these datasets test predictions do not appear prone to bias. Nonetheless this does not make them more reliable; on the contrary, their test predictions seem only enabled by the presence of symmetric and (in WN18) inverse relations. Table <ref type="table">3</ref> also displays that removing the test predictions featuring symmetric or inverse relations results in plummeting H@1 and MRR values, with a decrease usually between 50% and 75%.</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<monogr>
		<author>
			<persName><forename type="first">S</forename><surname>Auer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Bizer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Kobilarov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lehmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Cyganiak</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Ives</surname></persName>
		</author>
		<title level="m">Dbpedia: A nucleus for a web of open data</title>
				<imprint>
			<publisher>Springer</publisher>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
	<note>The semantic web</note>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">YAGO 4: A reason-able knowledge base</title>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">P</forename><surname>Tanon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Weikum</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">M</forename><surname>Suchanek</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2020">2020</date>
			<publisher>ESWC</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title level="m" type="main">Wikidata: a free collaborative knowledge base</title>
		<author>
			<persName><forename type="first">D</forename><surname>Vrandecic</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Krötzsch</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2014">2014</date>
			<publisher>CACM</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Collaboratively Built Semi-structured Content and Artificial Intelligence: The Story So Far</title>
		<author>
			<persName><forename type="first">E</forename><surname>Hovy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Navigli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">P</forename><surname>Ponzetto</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Artif. Intell</title>
		<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Semantic Parsing via Staged Query Graph Generation: Question Answering with Knowledge Base</title>
		<author>
			<persName><forename type="first">W</forename><surname>Yih</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Gao</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ACL</title>
				<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<author>
			<persName><forename type="first">A</forename><surname>Singhal</surname></persName>
		</author>
		<ptr target="https://www.blog.google/products/search/introducing-knowledge-graph-things-not/,blogpostintheOfficialGoogleBlog" />
		<title level="m">Introducing the knowledge graph: things, not strings</title>
				<imprint>
			<date type="published" when="2012">2012</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">Understand your world with bing</title>
		<author>
			<persName><forename type="first">R</forename><surname>Qian</surname></persName>
		</author>
		<ptr target="https://blogs.bing.com/search/2013/03/21/understand-your-world-with-bing/,blogpostinBingBlogs" />
		<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">Building a broad knowledge graph for products</title>
		<author>
			<persName><forename type="first">X</forename><forename type="middle">L</forename><surname>Dong</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2019">2019</date>
			<publisher>ICDE</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">Cracking the code on conversational commerce</title>
		<author>
			<persName><forename type="first">R</forename><surname>Pittman</surname></persName>
		</author>
		<ptr target="https://www.ebayinc.com/stories/news/cracking-the-code-on-conversational-commerce/,blogpostinEbayInc.Stories" />
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">Introducing graph search beta</title>
		<author>
			<persName><forename type="first">T</forename><surname>Stocky</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Rasmussen</surname></persName>
		</author>
		<ptr target="https://newsroom.fb.com/news/2013/01/introducing-graph-search-beta/,blogpostinFacebookNewsroom" />
		<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<title level="m" type="main">Building the linkedin knowledge graph</title>
		<author>
			<persName><forename type="first">Q</forename><surname>He</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B.-C</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Agarwal</surname></persName>
		</author>
		<ptr target="https://engineering.linkedin.com/blog/2016/10/building-the-linkedin-knowledge-graph/,blogpostinLinkedInBlog" />
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">Knowledge base completion via search-based question answering</title>
		<author>
			<persName><forename type="first">R</forename><surname>West</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Gabrilovich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Murphy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Lin</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2014">2014</date>
			<publisher>WWW</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<monogr>
		<title level="m" type="main">Realistic Re-evaluation of Knowledge Graph Completion Methods: An Experimental Study</title>
		<author>
			<persName><forename type="first">F</forename><surname>Akrami</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">S</forename><surname>Saeef</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Hu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Li</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2020">2020</date>
			<publisher>SIGMOD</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">Knowledge graph embedding for link prediction: A comparative analysis</title>
		<author>
			<persName><forename type="first">A</forename><surname>Rossi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Barbosa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Firmani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Matinata</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Merialdo</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2021">2021</date>
			<publisher>ACM TKDD</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">A survey on knowledge graph embeddings for link prediction</title>
		<author>
			<persName><forename type="first">M</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Qiu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Symmetry</title>
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<title level="m" type="main">Translating embeddings for modeling multi-relational data</title>
		<author>
			<persName><forename type="first">A</forename><surname>Bordes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Usunier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Garcia-Duran</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Weston</surname></persName>
		</author>
		<author>
			<persName><forename type="first">O</forename><surname>Yakhnenko</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2013">2013</date>
			<publisher>NIPS</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">Freebase: a collaboratively created graph database for structuring human knowledge</title>
		<author>
			<persName><forename type="first">K</forename><surname>Bollacker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Evans</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Paritosh</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Sturge</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Taylor</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2008">2008</date>
			<publisher>SIGMOD</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<title level="m" type="main">Observed versus latent features for knowledge base and text inference</title>
		<author>
			<persName><forename type="first">K</forename><surname>Toutanova</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Chen</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2015">2015</date>
			<publisher>CVSC</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m" type="main">Convolutional 2d knowledge graph embeddings</title>
		<author>
			<persName><forename type="first">T</forename><surname>Dettmers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Minervini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Stenetorp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Riedel</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2018">2018</date>
			<publisher>AAAI</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<monogr>
		<title level="m" type="main">Knowledge Graph Embeddings: Are Relation-Learning Models Learning Relations?</title>
		<author>
			<persName><forename type="first">A</forename><surname>Rossi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Matinata</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2020">2020</date>
			<pubPlace>PIE</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<author>
			<persName><forename type="first">J</forename><surname>Fisher</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Palfrey</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Christodoulopoulos</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Mittal</surname></persName>
		</author>
		<idno type="arXiv">arXiv:1912.02761</idno>
		<title level="m">Measuring social bias in knowledge graph embeddings</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Bias in knowledge graph embeddings</title>
		<author>
			<persName><forename type="first">S</forename><surname>Bourli</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Pitoura</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">ASONAM, IEEE</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Sample selection bias as a specification error</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">J</forename><surname>Heckman</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Econometrica</title>
		<imprint>
			<date type="published" when="1979">1979</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<monogr>
		<title level="m" type="main">Wordnet: a lexical database for english</title>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">A</forename><surname>Miller</surname></persName>
		</author>
		<imprint>
			<date type="published" when="1995">1995</date>
			<publisher>CACM</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<monogr>
		<title level="m" type="main">YAGO3: A knowledge base from multilingual wikipedias</title>
		<author>
			<persName><forename type="first">F</forename><surname>Mahdisoltani</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Biega</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><forename type="middle">M</forename><surname>Suchanek</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2015">2015</date>
			<publisher>CIDR</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<monogr>
		<title level="m" type="main">Complex Embeddings for Simple Link Prediction</title>
		<author>
			<persName><forename type="first">T</forename><surname>Trouillon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Welbl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Riedel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">É</forename><surname>Gaussier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Bouchard</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2016">2016</date>
			<publisher>ICML</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">TuckER: Tensor Factorization for Knowledge Graph Completion</title>
		<author>
			<persName><forename type="first">I</forename><surname>Balazevic</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Allen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">M</forename><surname>Hospedales</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">EMNLP -IJCNLP</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<monogr>
		<title level="m" type="main">Interaction embeddings for prediction and explanation in knowledge graphs</title>
		<author>
			<persName><forename type="first">W</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Paudel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Bernstein</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Chen</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2019">2019</date>
			<publisher>WSDM</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b28">
	<monogr>
		<title level="m" type="main">Learning Hierarchy-Aware Knowledge Graph Embeddings for Link Prediction</title>
		<author>
			<persName><forename type="first">Z</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Cai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Wang</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2020">2020</date>
			<publisher>AAAI</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b29">
	<monogr>
		<title level="m" type="main">InteractE: Improving Convolution-based Knowledge Graph Embeddings by Increasing Feature Interactions</title>
		<author>
			<persName><forename type="first">S</forename><surname>Vashishth</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Sanyal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">V</forename><surname>Nitin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><surname>Agrawal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">P</forename><surname>Talukdar</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2020">2020</date>
			<publisher>AAAI</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b30">
	<monogr>
		<title level="m" type="main">Learning to Exploit Long-term Relational Dependencies in Knowledge Graphs</title>
		<author>
			<persName><forename type="first">L</forename><surname>Guo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Hu</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2019">2019</date>
			<publisher>ICML</publisher>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b31">
	<analytic>
		<title level="a" type="main">A Capsule Network-based Embedding Model for Knowledge Graph Completion and Search Personalization</title>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">Q</forename><surname>Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Vu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">D</forename><surname>Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">Q</forename><surname>Nguyen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">Q</forename><surname>Phung</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">NAACL-HLT</title>
				<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b32">
	<monogr>
		<author>
			<persName><forename type="first">C</forename><surname>Meilicke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">W</forename><surname>Chekol</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Fink</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Stuckenschmidt</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2004.04412</idno>
		<title level="m">Reinforced Anytime Bottom Up Rule Learning for Knowledge Graph Completion</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
</biblStruct>

<biblStruct xml:id="b33">
	<analytic>
		<title level="a" type="main">Link prediction of weighted triples for knowledge graph completion within the scholarly domain</title>
		<author>
			<persName><forename type="first">M</forename><surname>Nayyeri</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">M</forename><surname>Cil</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Vahdati</surname></persName>
		</author>
		<author>
			<persName><forename type="first">F</forename><surname>Osborne</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Kravchenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Angioni</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">A</forename><surname>Salatino</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">R</forename><surname>Recupero</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Motta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lehmann</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Access</title>
		<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b34">
	<analytic>
		<title level="a" type="main">Popularity agnostic evaluation of knowledge graph embeddings</title>
		<author>
			<persName><forename type="first">A</forename><surname>Mohamed</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Parambath</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Kaoudi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Aboulnaga</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">UAI, PMLR</title>
				<imprint>
			<date type="published" when="2021">2021</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
