<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Optimal Transport Methods for Aligning Knowledge Graph Triples with Natural Language in Unsupervised Settings</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Alexander</forename><surname>Kalinowski</surname></persName>
							<email>ajk437@drexel.edu</email>
							<affiliation key="aff0">
								<orgName type="institution">Drexel University</orgName>
								<address>
									<postCode>19104</postCode>
									<settlement>Philadelphia</settlement>
									<region>PA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Optimal Transport Methods for Aligning Knowledge Graph Triples with Natural Language in Unsupervised Settings</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">0B1F7BAF2F545C722039A9513773B14B</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T14:31+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Knowledge representation</term>
					<term>Automated metadata generation</term>
					<term>Embeddings</term>
					<term>Optimal transport</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Frameworks for aligning embeddings of text and embeddings of knowledge graphs (KG) have been used for generating mappings for test-to-text alignment and KG-to-KG alignment, but little has been done for alignment between these two domains. In this dissertation proposal, I aim to create a framework for KG-to-text alignment that utilizes little to no training data to learn these correspondences. Additionally, motivated by the semantic geometries of these embedding spaces, I propose a new line of research into generating explicit embeddings of triples from a knowledge graph.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">Importance</head><p>Knowledge graphs (KGs) and ontologies form the computational backbone of the modern Semantic Web, curated by taxonomists and ontologists in conjunction with domain subject matter experts. Collaboration between these parties is a bottleneck in large-scale organizations due to coordination of people and sourcing of relevant information and terminologies for ontologists to massage into an enterprise-wide standard. This bottleneck is felt both in developing ontological terminologies and populating the knowledge graph with facts and assertions about the domain being described.</p><p>Industrial terminologies are buried deep in policy documents or technical white papers, making the job of the ontologist one of synthesizing these documents into a conscise, machine-readable set of interlinked terminologies (T-Box). For large-scale knowledge graphs, such as those leveraged in popular search engines, the information scales past terminologies and additionally covers facts or assertions (A-Box) about the objects described in the graph. Validating the accuracy of assertions in the knowledge graph is critical for auditing the trustworthiness of those claims, which is especially relevant in highly regulated industries such as banking or pharmatecuticals. As facts (in the form of s, p, o triples) are added to a knowledge graph, either through automated methods such as link prediction techniques or human generated annotations, there is an additional opportunity to enrich the knowledge graph with metadata about these triples, such as a source of evidence from a text document. Developing methodologies for linking the T-Box and A-Box to textual evidence will provide a set of tools to allow ontologists and knowledge graph developers to expedite their work while ensuring the highest degree of accuracy and auditability, and thus is the focus of this work.</p><p>For example, an ontologist working in the financial services domain may wish to develop a standardized definition of a fixed-float interest rate swap. They may begin by inheriting the structure of previously defined terminologies, namely those related to interest rates and swaps, deriving the triples f ixed f loat interest rate swap, has type, swap contract , f ixed f loat interest rate swap, has leg, f ixed interest rate and f ixed f loat interest rate swap, has leg, f loating interest rate .</p><p>However, without a background in financial terminology, they may miss the fact that a fixed-float interest rate swap is more commonly refered to as a vanilla interest rate swap. This fact can be inferenced from a variety of textual sources, such as a sentence like 'A vanilla interest rate swap allows two counterparties to hedge against interest rate volatility by trading a floating rate for a fixed rate.', given the set of seed triples the ontologist has developed, helping to expand the coverage of the knowledge graph.</p><p>Of critical importance in the development of a text to KG methodology is its ability to rapidly adapt to new source ontologies and data domains. Additionally, such a system should not be bottlenecked by reliance on abundant training data, a point of failure for many ML projects. This motivates the use of unsupervised techniques for knowledge representation, and I propose an exploration of cross-domain optimal transport between embeddings of natural language and embeddings of a knowledge graph. Given this motivation, I present the following formalism.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">Problem Statement</head><p>Suppose S = {s 1 , s 2 , . . . , s n } is a set of sentences and T = {t 1 , t 2 , . . . , t n } is a set of knowledge graph triples, each triple of the form t i = s, p, o . Next, define two functions f and g that create low-dimensional (latent) representations of each sentence and triple, respectively, such that f (s i ) = e si ∈ R n (the source space) and g(t j ) = e tj ∈ R m (the target space) where n |S|, m |T |. In order to link triples to their most relevant sentences, both for validating terminology and assertions, we seek a mapping function Ψ : R n → R m to transport one set to another with minimal loss, i.e. Ψ (T ) ≈ S, such that for new triples t i / ∈ T we can find a supporting sentence representation Ψ (t i ) ≈ s j / ∈ S in a large-scale text corpus C that reflects the same semantic information. Learning such a Ψ should not be taxed by reliance on an abundance of paired labeled data points, i.e. a set of labels L = {(t k , s k ), . . . , (t l , s l )}. Instead, I assume no such set exists at training time, framing the learning of Ψ as an unsupervised task.</p><p>The desire to limit the reliance on paired labeled data points helps inform the potential choices of Ψ . Specifically, methods from optimal transport (OT) theory lend themselves well to this problem by exploiting the structure of each embedding space using pairwise distance metrics rather than relying on data with paired labels. Instead, couplings of objects from each respective space are inferred through probabilistic transport maps, shifting the focus from gathering labeled sentence-triple pairs for supervised learning to refining the representations of these objects in their respective latent spaces. Optimal transport also provides a probabilistic framework for mapping assignments; the Kantorivitch relaxation of Monge's original statement admits a solution where the mass at any point in the source space can be dispatched to several locations in the target space <ref type="bibr" target="#b17">[18]</ref>, fitting the problem setting as a single KG triple may admit infinitely innumerable sentence representations. The probabilistic and rigorous mathematical approach of OT add to the understandability of results in opposition to black-box models such as generative adversarial networks (GANs).</p><p>One drawback of OT techniques is the requirement of a cost matrix defined between the spaces X and Y. Defining such a cost matrix requires labeled data points beween the two spaces, although a weaker assumption can be used to avoid this by defining two inter-domain distance matrices D ∈ R n×n and D ∈ R m×m . Such matrices can then be aligned through the following formulation of the Gromov-Wasserstein problem</p><formula xml:id="formula_0">GW ((a, D), (b, D )) 2 = min P ∈U (a,b) E D,D (P ) E D,D (P ) = Σ i,j,i ,j |D i,i − D j,j | 2 P i,j P i ,j</formula><p>.</p><p>Using this formalism, the problem of interest can then be framed as follows:</p><p>Problem Statement: What are the optimal choices of embedding functions f and g to establish distance matrices D and D such that 1. for pairs of triples (t i , t i ) ∈ T that contain some notion of semantic similarity,</p><formula xml:id="formula_1">D i,i = d(f (t i ), f (t i )</formula><p>) is minimized 2. while simultaneously minimizing D j,j = d(g(s j ), g(s j )) for semantically similar sentences (s j , s j ) ∈ S for some distance metric d, thus allowing optimal couplings between t i and s i to be established via the Gromov-Wasserstein distance?</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">Research Questions, Hypotheses and Research Plan</head><p>Prior research in this area shows that optimal transport of embedding spaces has been successful within a given data domain (i.e. unsupervised alignment of word embeddings <ref type="bibr" target="#b1">[2]</ref>, unsupervised alignment of knowledge graph entities <ref type="bibr" target="#b16">[17]</ref>). My research questions seek to extend these methodologies to the cross-domain task in order to align embeddings of knowledge graph triples with embeddings of semantically related sentences.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>RQ-I:</head><p>Is there an accurate, unsupervised technique for aligning a set of knowledge graph triples to a set of semantically similar sentences?</p><p>H-I: Optimal transport methods provide a mathematical framework for unsupervised alignment based on intra-domain pairwise similarities. Successful application of OT is dependent on how similarities of like-objects are represented in each respective space.</p><p>To accomplish this, properties of each embedding space that make them useful for such alignment must first be established. In my prior work <ref type="bibr" target="#b8">[9]</ref>, these properties are explored for sentence embeddings while keeping the knowledge graph embeddings fixed as the concatenation of head, tail and relation vectors generated by TransE. It remains to be seen how changing the structure of knowledge graph embeddings, via changing the algorithm selection (such as using more expressive models like ComplEx <ref type="bibr" target="#b19">[20]</ref> or ConvE <ref type="bibr" target="#b6">[7]</ref>), incorporating additional information such as literals <ref type="bibr" target="#b10">[11]</ref> or changing the way entity and relation vectors are combined to represent a triple, help or hurt the ability to generate high-quality alignments, motivating the next research question.</p><p>RQ-II: How can current knowledge graph embedding methods be extended past representing entities and relations as separate objects and instead focus on embedding triples as the target objects?</p><p>As the majority of current methods focus almost exclusively on the link prediction task, these methods may not be well-suited for establishing embeddings of triples, leading to the following hypothesis.</p><p>H-II: Triple embeddings built from aggregations of entity and relation embeddings do not sufficiently encode the underlying semantics of such triples.</p><p>Building upon the work of <ref type="bibr" target="#b7">[8]</ref>, treating triples as walks on the knowledge graph and weighting the strength of each relationship may help to create a semantic embedding space that will assist in alignment. The following section details how I will approach measuring the amount of semantic information captured by these methods.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">Approach and Evaluation</head><p>Motivated by work on word embedding regularities <ref type="bibr" target="#b13">[14]</ref>, I wish to probe both sentence and KG embedding spaces generated by a variety of embedding algorithms and measure the degree to which they exhibit an underlying structure that can be leveraged for aligning these resources. Lurking beneath the above research questions is the fuzzily-defined notion of "semantic similarity," but metrics exist to make this quantification concrete. These metrics are used to define how well semantic similarity is encoded in the latent representations of both triples and sentences, and they are important to capture with the goal of defining optimal pair-wise distance matrices D and D in mind. To formalize the notion of structure, I introduce a definition of clusterability, following the work of <ref type="bibr" target="#b0">[1]</ref>. For some dataset X ∈ R n , a description of the clusterability of X is a function c : X → v where v ∈ R is a real value. Here, v is a measure of how strong a clustering presence is in the underlying set X.</p><p>To test the clusterability hypothesis, I use the spatial histogram (SpatHist) approach to measure the clusterability of each space <ref type="bibr" target="#b0">[1]</ref>. The SpatHist approach compares the data binned in all d-dimensions to samples randomly generated in the same d-dimensional space. As many of these bins may end up empty in high-dimensional embedding spaces, I perform principal component analysis (PCA) to project down to the two most informative dimensions, split the data into n equal-width bins, and compute the empirical joint probability mass function (EPMF). The same is then done for 500 sets of uniformly generated points with the same feature dimensionality, and the differences are compared using the Kullback-Leibler (KL) divergence -higher KL divergence indicates more clusterability. I report the mean and standard deviation of each of these experiments as my final estimates of clusterability. Additionally, I apply the Hopkins test of uniformity <ref type="bibr" target="#b11">[12]</ref>. As the Hopkins test statistic tends to zero, the underlying data exhibits less uniformity, indicating that clustering may be a good way of exploring the data in an unsupervised way. As the test statistic increases, the data tends to be more uniformly distributed, exhibiting less of the structure I seek to exploit.</p><p>For evaluating the quality of learned alignments, I follow in the tradition of knowledge graph embedding literature <ref type="bibr" target="#b20">[21]</ref> and evaluate these results for the Hits@5 and Hits@10 metrics. Analyzing the results of the top 5 and top 10 closest matches allows for a nearest neighborhood analysis of each aligned embbedding instance, helping to pinpoint areas for future improvement, such as the mitigation of the influence of hubs <ref type="bibr" target="#b12">[13]</ref> .</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">Preliminary Results</head><p>To establish a baseline for this task, prior work <ref type="bibr" target="#b8">[9]</ref> tested the ability for a lowcapacity linear model to learn a mapping between sentence and knowledge graph representations. The purpose of this work is in evaluating sentence representations, measuring the extent to which they are able to create structure in the low-dimensional embedding spaces by evaluating how well they cluster together around their semantic content, in this case the expression of a particular relationship. Findings on the clustering capacity of a selection of sentence embedding methods are reproduced below.</p><p>The results demonstrate the dramatic differences in the efficacy of sentence embedding methodologies. In particular, the geometrically motivated GEM algorithm vastly outperforms all others in terms of semantic clusterability, especially those using more complex deep neural models. In addition, the GEM algorithm outperforms all others in terms of Hits@5 and Hits@10 when performing a simple linear map for alignment (Linear@5,10), and all alignments show improvement when replacing the linear alignment with optimal transport techniques (OT@5,10). Utilizing these results gives a clue as to how to build knowledge graph triple embeddings: by focusing on the novelty of each predicate and the entities involved, they can be "pushed" into respective areas of the low-dimensional embedding space, leading to increased cluster cohesion and higher within cluster semantic similarity. Additional insights and recommended improvements are presented in <ref type="bibr" target="#b8">[9]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Model</head><p>Dim. Linear@5 Linear@10 OT@5 OT@ </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.1">Natural Language Representation Alignment</head><p>Regression models for word-to-word alignment were first proposed by <ref type="bibr" target="#b14">[15]</ref> as a means of capturing geometric patterns between embeddings across embedding spaces. Inconsistencies in this approach were noted by <ref type="bibr" target="#b21">[22]</ref> who in turn modified the regression process to add unit length normalization and constrain map to be orthogonal. Applications of pre-processing and orthogonal constraints spurred further research into ways to manipulate the source and target embedding spaces to further express their geometric structures in <ref type="bibr" target="#b2">[3]</ref> and <ref type="bibr" target="#b3">[4]</ref>. Building 'pseudodictionaries' as a means of reducing the amount of necessary training data is suggested in <ref type="bibr" target="#b18">[19]</ref>. The work of <ref type="bibr" target="#b15">[16]</ref> further explores iterative learning, alternating between supervised alignment and unsupervised distribution matching, as well as introducing novel metrics to assess the orthogonality assumptions used in supervised approaches. A key approach for unsupervised learning is described in <ref type="bibr" target="#b5">[6]</ref> where the authors propose leveraging an adversarial learning paradigm. While the adversarial method directly leverages word frequencies, an alternative unsupervised method in <ref type="bibr" target="#b4">[5]</ref> captures these patterns by analyzing the similarity distributions of the word vectors themselves. Using the Gromov-Wasserstein distance, <ref type="bibr" target="#b1">[2]</ref> transform the alignment problem to one of finding an optimal transport from source X to target Z.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.2">Graph Representation Alignment</head><p>The majority of research in the area of knowledge graph embeddings focus on one specific task, namely knowledge graph completion, which seeks to make predictions of the following form: given s, p, ? , make the best prediction for an object o such that the triple is a valid one in the context of the greater graph. The state-of-the-art methods in this space leverage knowledge graph embeddings, low-dimensional representations of the entities and relations between them as vectors. The majority of research in this area focuses on representing the nodes, or entities, of the graph, with considerable less emphasis on how the relations are represented, and representations of the entire s, p, o triple are rarely considered.</p><p>Recent research into representing the entire triple is presented in <ref type="bibr" target="#b7">[8]</ref>, yet the results are limited to explorations in clustering and recommendation systems. This work can be extended to further use cases and eventually tied in with alignment methods to link knowledge graphs to text documents.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7">Reflection and Future Work</head><p>Based on the initial results presented herein, is clearly room for improvement in methodologies for representation learning of triples in a knowledge graph. Additionally, the alignment of cross-domain representations -those spanning both text and knowledge graph, is not currently well explored. Exploration in this area can provide a brigde between the two respective data realms and provide tooling for unsupervised automation of ontology and knowledge graph development. Future work will involve measuring the clusterability of multiple existing knowledge graph embedding algorithms, evaluating their efficacy in alignments with sentence embeddings, and proposing new, semantically grounded approaches to embedding triples as singular objects.</p></div>			<note xmlns="http://www.tei-c.org/ns/1.0" place="foot" xml:id="foot_0">Proceedings of the Doctoral Consortium at ISWC 2021 -ISWC-DC 2021</note>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">To cluster, or not to cluster: An analysis of clusterability methods</title>
		<author>
			<persName><forename type="first">A</forename><surname>Adolfsson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ackerman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">C</forename><surname>Brownstein</surname></persName>
		</author>
		<idno type="DOI">10.1016/j.patcog.2018.10.026</idno>
		<ptr target="http://dx.doi.org/10.1016/j.patcog.2018.10.026" />
	</analytic>
	<monogr>
		<title level="j">Pattern Recognition</title>
		<imprint>
			<biblScope unit="volume">88</biblScope>
			<biblScope unit="page" from="13" to="26" />
			<date type="published" when="2019-04">Apr 2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title level="m" type="main">Gromov-wasserstein alignment of word embedding spaces</title>
		<author>
			<persName><forename type="first">D</forename><surname>Alvarez-Melis</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">S</forename><surname>Jaakkola</surname></persName>
		</author>
		<idno>CoRR abs/1809.00013</idno>
		<ptr target="http://arxiv.org/abs/1809.00013" />
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title level="m" type="main">Learning principled bilingual mappings of word embeddings while preserving monolingual invariance</title>
		<author>
			<persName><forename type="first">M</forename><surname>Artetxe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Labaka</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Agirre</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/D16-1250</idno>
		<ptr target="https://doi.org/10.18653/v1/D16-1250" />
		<imprint>
			<date type="published" when="2016-01">01. 2016</date>
			<biblScope unit="page" from="2289" to="2294" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">Generalizing and improving bilingual word embedding mappings with a multi-step framework of linear transformations</title>
		<author>
			<persName><forename type="first">M</forename><surname>Artetxe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Labaka</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Agirre</surname></persName>
		</author>
		<ptr target="https://www.aaai.org/ocs/index.php/AAAI/AAAI18/paper/view/16935" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence</title>
				<meeting>the Thirty-Second AAAI Conference on Artificial Intelligence</meeting>
		<imprint>
			<date type="published" when="2018-02">February 2018</date>
			<biblScope unit="page" from="5012" to="5019" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">A robust self-learning method for fully unsupervised cross-lingual mappings of word embeddings</title>
		<author>
			<persName><forename type="first">M</forename><surname>Artetxe</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Labaka</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Agirre</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Doctoral Consortium at ISWC 2021 -ISWC-DC 2021</title>
				<meeting>the Doctoral Consortium at ISWC 2021 -ISWC-DC 2021</meeting>
		<imprint>
			<date type="published" when="2018">2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<monogr>
		<title level="m" type="main">Word translation without parallel data</title>
		<author>
			<persName><forename type="first">A</forename><surname>Conneau</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Lample</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Ranzato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Denoyer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">H</forename><surname>Jégou</surname></persName>
		</author>
		<idno>CoRR abs/1710.04087</idno>
		<ptr target="http://arxiv.org/abs/1710.04087" />
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<author>
			<persName><forename type="first">T</forename><surname>Dettmers</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Minervini</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Stenetorp</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Riedel</surname></persName>
		</author>
		<title level="m">Convolutional 2d knowledge graph embeddings</title>
				<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<monogr>
		<title level="m" type="main">Triple2vec: Learning triple embeddings from knowledge graphs</title>
		<author>
			<persName><forename type="first">V</forename><surname>Fionda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Pirrò</surname></persName>
		</author>
		<idno>CoRR abs/1905.11691</idno>
		<ptr target="http://arxiv.org/abs/1905.11691" />
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">A comparative study on structural and semantic properties of sentence embeddings</title>
		<author>
			<persName><forename type="first">A</forename><surname>Kalinowski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>An</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<monogr>
		<title level="m" type="main">A survey of embedding space alignment methods for language and knowledge graphs</title>
		<author>
			<persName><forename type="first">A</forename><surname>Kalinowski</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>An</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<monogr>
		<title level="m" type="main">Incorporating literals into knowledge graph embeddings</title>
		<author>
			<persName><forename type="first">A</forename><surname>Kristiadi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Khan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Lukovnikov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Lehmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Fischer</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2019">2019</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">New index for clustering tendency and its application to chemical problems</title>
		<author>
			<persName><forename type="first">R</forename><forename type="middle">G</forename><surname>Lawson</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><forename type="middle">C</forename><surname>Jurs</surname></persName>
		</author>
		<idno type="DOI">10.1021/ci00065a010</idno>
		<ptr target="https://pubs.acs.org/doi/abs/10.1021/ci00065a010" />
	</analytic>
	<monogr>
		<title level="j">Journal of Chemical Information and Computer Sciences</title>
		<imprint>
			<biblScope unit="volume">30</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="36" to="41" />
			<date type="published" when="1990">1990</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Hubness and pollution: Delving into cross-space mapping for zero-shot learning</title>
		<author>
			<persName><forename type="first">A</forename><surname>Lazaridou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Dinu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Baroni</surname></persName>
		</author>
		<idno type="DOI">10.3115/v1/P15-1027</idno>
		<ptr target="https://www.aclweb.org/anthology/P15-1027" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics)</title>
				<meeting>the 53rd Annual Meeting of the Association for Computational Linguistics)<address><addrLine>Beijing, China</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="270" to="280" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<monogr>
		<title level="m" type="main">Exploiting similarities among languages for machine translation</title>
		<author>
			<persName><forename type="first">T</forename><surname>Mikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><forename type="middle">V</forename><surname>Le</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sutskever</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<monogr>
		<title level="m" type="main">Exploiting similarities among languages for machine translation</title>
		<author>
			<persName><forename type="first">T</forename><surname>Mikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Q</forename><forename type="middle">V</forename><surname>Le</surname></persName>
		</author>
		<author>
			<persName><forename type="first">I</forename><surname>Sutskever</surname></persName>
		</author>
		<idno>CoRR abs/1309.4168</idno>
		<ptr target="http://arxiv.org/abs/1309.4168" />
		<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Bilingual lexicon induction with semi-supervision in non-isometric embedding spaces</title>
		<author>
			<persName><forename type="first">B</forename><surname>Patra</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">R A</forename><surname>Moniz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Garg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">R</forename><surname>Gormley</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Neubig</surname></persName>
		</author>
		<ptr target="https://www.aclweb.org/anthology/P19-1018" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics</title>
				<meeting>the 57th Annual Meeting of the Association for Computational Linguistics<address><addrLine>Florence, Italy</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="184" to="193" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">Improving cross-lingual entity alignment via optimal transport</title>
		<author>
			<persName><forename type="first">S</forename><surname>Pei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Zhang</surname></persName>
		</author>
		<idno type="DOI">10.24963/ijcai.2019/448</idno>
		<ptr target="https://doi.org/10.24963/ijcai.2019/448" />
		<imprint>
			<date type="published" when="2019-08">08. 2019</date>
			<biblScope unit="page" from="3231" to="3237" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<monogr>
		<author>
			<persName><forename type="first">G</forename><surname>Peyré</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Cuturi</surname></persName>
		</author>
		<title level="m">Computational optimal transport</title>
				<imprint>
			<date type="published" when="2020">2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m" type="main">Offline bilingual word vectors, orthogonal transformations and the inverted softmax</title>
		<author>
			<persName><forename type="first">S</forename><forename type="middle">L</forename><surname>Smith</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><forename type="middle">H P</forename><surname>Turban</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Hamblin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">Y</forename><surname>Hammerla</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Complex embeddings for simple link prediction</title>
		<author>
			<persName><forename type="first">T</forename><surname>Trouillon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Welbl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Riedel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">E</forename><surname>Gaussier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Bouchard</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">International Conference on Machine Learning (ICML)</title>
				<imprint>
			<date type="published" when="2016">2016</date>
			<biblScope unit="volume">48</biblScope>
			<biblScope unit="page" from="2071" to="2080" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">Knowledge graph embedding: A survey of approaches and applications</title>
		<author>
			<persName><forename type="first">Q</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Z</forename><surname>Mao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><surname>Guo</surname></persName>
		</author>
		<idno type="DOI">10.1109/TKDE.2017.2754499</idno>
		<ptr target="https://doi.org/10.1109/TKDE.2017.2754499" />
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Knowledge and Data Engineering PP</title>
		<imprint>
			<biblScope unit="page" from="1" to="1" />
			<date type="published" when="2017-09">09 2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">Normalized word embedding and orthogonal transform for bilingual word translation</title>
		<author>
			<persName><forename type="first">C</forename><surname>Xing</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Y</forename><surname>Lin</surname></persName>
		</author>
		<ptr target="ISWC-DC2021" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2015 Conference of the North American Chapter of the Association for Computational Linguistics</title>
				<meeting>the 2015 Conference of the North American Chapter of the Association for Computational Linguistics<address><addrLine>Denver, Colorado</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2015">2015. 2021 -</date>
			<biblScope unit="page" from="1006" to="1011" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
