<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Enriched Network Embeddings for News Recommendation</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author role="corresp">
							<persName><forename type="first">Janu</forename><surname>Verma</surname></persName>
							<email>j.verma5@gmail.com</email>
							<affiliation key="aff0">
								<orgName type="institution">Hike Messenger New Delhi</orgName>
								<address>
									<settlement>Delhi</settlement>
									<country key="IN">India</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Enriched Network Embeddings for News Recommendation</title>
					</analytic>
					<monogr>
						<imprint>
							<date/>
						</imprint>
					</monogr>
					<idno type="MD5">81E2C15EEC52F264CBC332375CADD1CC</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2023-03-24T05:27+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>News recommendation</term>
					<term>network embeddings</term>
					<term>NLP</term>
					<term>Named Entities</term>
					<term>Topic Modeling</term>
					<term>Collaborative Filtering</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>News aggregators collects content from various sources and presents them in one website or mobile application for easy access. A key challenge for the news applications is to help users discover relevant articles. Both the user experience and the key metrics depend on the high-quality personalized recommendations. However, building a news recommendation presents a set of challenges due the large number of articles being published every hour, the surge and decline in the popularity of news, and critical nature of recency etc. In this paper, we present a graph-based news recommendation model which is deployed on a real-world news application. Our system is a hybrid of collaborative-filtering and the content-based filtering. We enrich the user-article interaction graph by adding new nodes corresponding to the named entities extracted from the contents of the articles. The random walk based graph embeddings are used to learn latent representation for users, articles and named entities in the same space. We evaluate the learned embeddings via a multi-class classification of news articles into high-level categories. We propose a recommendation system based on the binary classification problem which takes as input a combination of the user, item and entity embeddings and computes the probability of the user clicking on the article. We perform experiments to show the superiority of our model to the previous system.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1">INTRODUCTION</head><p>News reading on the mobile devices has become very popular in recent years owing to the availability of tremendous amount of data coming from millions of sources across the world. As per a 2017 survey by Pew Research <ref type="bibr">[8]</ref>, about 85% of US adults read news on a mobile device. News applications like Google News, Bing News, Flipboard, Pocket etc. collect news from various sources and provide readers with news from around the world in an aggregated user-interface. The sheer volume of available articles can be overwhelming to the readers. A key problem that news aggregation applications struggle with is to help users discover articles that are most interesting to them.</p><p>Recommendation system aims to capture users preferences and interests so that relevant information is shown to them. Domains like online shopping (e.g. Amazon), movies (e.g. Netflix), music (e.g. Pandora) have seen great success of recommendation systems. Two popular approaches of recommendation systems are collaborative filtering and contentbased filtering which form the basis for major recommendation systems. In collaborative-filtering, news articles are recommended based on the reading history of users with similar preferences. This is a very popular method which has the advantage of being domain free. A major drawback of collaborative-filtering is its inability to handle new content i.e. a breaking news that has not yet seen much traction, this is called the item cold-start problem.</p><p>A solution of the problem of fresh content would be to build user profiles comprising of genuine interests and use them to make news recommendations. This is called content-filtering where the contents of the articles are analyzed to extract topics of user's interest. If a user's past reading preferences are known, new articles can be recommended based on their similarity to the previously read. Content-based filtering requires sufficient reading history of users to be able to build a strong profile of user's interests. For users with little history, i.e. user cold-start, this becomes problematic. Collaborativefiltering relies on the fact that there are always users who read some news, and these users may serve as a basis to help to predict the interests of the long-tail users. Thus, collaborativefiltering and content-based filtering are complementary to each other.</p><p>The user preferences are not straightforward-a user might want to read an article even if she is not interested in the topic but finds the particular story relevant. For example, wanting to read news about World Cup even if no general interest in Sports. This requires a carefully designed contentfiltering model at a proper level of granularity. Furthermore, not all users are equal to each other, and the collaborative filtering method may not account for the individual variability between users. Highly read topics are recommended to most of the users, even if some of the them have no interest in these topics. For example, entertainment and lifestyle articles are most popular and they get reflected in the recommendations for a lot of users via other seemingly similar users.</p><p>Thus, news recommendation presents challenges that do not exist in other domains. For example, the recency and the popularity of news articles can change drastically with time. Another compounding factor is the influx of a large number of new articles every hour.</p><p>to vectors in a low-dimensional space such that the structural attributes of the graph translate to the geometrical properties in the embedding space. The user-item interaction data can be defined as a bipartite graph with user nodes and item nodes. In collaborative filtering, the adjacency matrix of this bipartite graph is used to learn similarity of users and items <ref type="bibr" target="#b8">[10]</ref>. In the current work, we enrich the user-item graph by adding new nodes corresponding to the named entities -person, place, organization, event -extracted from the contents of the articles. For instance, if an article is about 2016 Presidential elections, we might add entity nodes like -Donald Trump, 2016 Elections, Republican Party etc. Also added are the edges from the item nodes to the entity nodes. The enriched graph has three types of nodes -user, item, entity, and 2 types of edges -user-item and item-entity.</p><p>To learn node embeddings, truncated random walks (biased or unbiased) of fixed length are generated emanating from each node <ref type="bibr" target="#b4">[5]</ref>. These random walk sequences can be thought of as sentences in an artificial language. Using the Skip-gram <ref type="bibr" target="#b6">[7]</ref> model where words in a corpus of sentences in a natural language can be mapped to a low-dimensional space, we obtain dense representations for nodes. The embeddings aim to capture contextual similarity of the nodes i.e. nodes cooccurring on a fixed window on a random walk are mapped to nearby points. In collaborative-filtering nodes which cooccur in the adjacency list (direct connections) of a node are mapped to nearby points <ref type="bibr" target="#b8">[10]</ref>. Thus, graph embeddings provide an extension of the collaborative-filtering and has been shown to perform better.</p><p>Using graph embedding, we map all nodes-users, items and entities-to the same space. In this space, we can compute similarity of a user to an item and also the user-entity similarity. This allows us to suggest news articles based on a user's affinity towards the entities which are at a much lower level than topical affinity in content-based filtering. This equips us to handle the situation where e.g. a consumer is interested in reading the news about elections even if she is not generally interested in politics. The entities are also more interpretable than abstract topics.</p><p>We provide an evaluation of the learned representation by studying its efficacy in multi-class classification of the article nodes into 8 pre-defined high-level categories e.g. Politics, Entertainment, Sports etc. A simple linear classifier trained only on the node embeddings of 60% of the article nodes (without using any content information explicitly) labeled with their respective categories, and evaluated on the remaining 40% provides 0.901 AUC. We also cluster embeddings of the entity nodes and qualitatively evaluate the results. Finally, we evaluate the system for article recommendation by computing Precision@k for k values ranging from 1 to 5.</p><p>Concretely, we make following contributions:</p><p>• Provide a news recommendation system based on graph embeddings that is a hybrid of collaborative and contentbased filtering.</p><p>• Learn embeddings for the enriched user-item graph that contains entity nodes capturing contents of the article in addition to the user and item nodes. • Evaluate the learned embeddings via multi-class article classification. • Build and evaluate the binary classification model for recommendation.</p><p>• Study the efficacy of our method for cold-start problem.</p><p>The remainder of the article is organized as follows: In Section 2, we provide a discussion of the related work on recommendation systems and network embeddings. Next, we explain graph embeddings for the bipartite and the enriched graphs and their utilization in the recommendation model in Section 3. Section 4 provides the analysis and evaluation of the proposed model. Finally, we conclude in Section 5.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2">RELATED WORK</head><p>In this section, we discuss some of the related work on news recommendation systems.</p><p>Collaborative Filtering: Collaborative filtering <ref type="bibr" target="#b8">[10]</ref> recommends articles which were clicked by users with similar reading history. Collaborative filtering has been applied to personalized news reading applications, such as GroupLens <ref type="bibr" target="#b9">[11]</ref> and in the initial version of Google News recommendation <ref type="bibr" target="#b7">[9]</ref>. There are two types of collaborative-filtering <ref type="bibr" target="#b8">[10]</ref> : em neighborhood model and latent factor model. In neighborhood model, the click history of users is used to compute the 'neighborhood' of a user and then articles are recommended to users from the articles clicked on by their neighbors. In the latent factor model, a latent representation of both users and articles in the same space, usually via a matrix factorization of the user-article interaction matrix. The latent representation can be interpreted as describing an article or a user in a 'concept' space which captures the factors e.g. topic of the article. The collaborative-filtering is employed when there is scarce click history available for users.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Content-based Filtering:</head><p>A Content-based recommendation system tries to recommend items similar to those a given user has liked in the past. Thus, it requires a notion of similarity of articles. There has been a lot of work in NLP to compute similarity between text documents e.g. tf-idf, word embeddings <ref type="bibr" target="#b6">[7]</ref> and doc2vec <ref type="bibr" target="#b10">[12]</ref>. This relies on sufficient click history to be able to build a genuine profile of user's interests. The content-based filtering has been applied to personalized news recommendations e.g. news reading on devices ( <ref type="bibr" target="#b11">[13]</ref> , <ref type="bibr" target="#b13">[15]</ref>) and web-based news aggregation services <ref type="bibr" target="#b12">[14]</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Hybrid Collaborative and Content-based Filtering:</head><p>Hybrid model which combine both collaborative-filtering and content-based filtering are more stable against the problems of any one of the approaches. This is accomplished by using both the user similarity based on historical information and the content similarity to make recommendation. Hybrid methods have been seen applications in news recommendation <ref type="bibr" target="#b14">[16]</ref> and <ref type="bibr" target="#b15">[17]</ref>.</p><p>Graph Embeddings: Our approach presented in this paper is based on graph embeddings. Instead of working with global summary attributes of the graph, there has been a lot of work recently to find a representation of the nodes that incorporates the local structural information <ref type="bibr" target="#b4">[5]</ref>, <ref type="bibr" target="#b2">[3]</ref>. The idea is to learn a mapping from the graph to a low-dimensional real vector space such that the structural attributes of the graph translate to the geometrical properties in the embedding space. Random walks of pre-decided length are generated starting at each node in the network to produce "sentences" of nodes, similar to sentences of words in a natural language. The Skip-gram algorithm devised by Mikolov et al. <ref type="bibr" target="#b6">[7]</ref> is used to obtain node embedding from the random walks, which are expected to capture the contextual properties of the network nodes : the nodes that occur in same context have similar vector embedding. For a survey on graph embeddings, see Cui et al. <ref type="bibr" target="#b5">[6]</ref>. Recently, random walks on graph and embeddings have been used in recommendation tasks e.g. Pixie <ref type="bibr" target="#b16">[18]</ref>, GraphSage <ref type="bibr" target="#b17">[19]</ref> etc.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3">PROPOSED METHOD</head><p>In this section, we describe our model for news recommendation which uses graph embeddings as feature learning. News aggregation applications gather data from various sources and present the results as a list to the users. The recommendation system attempts to rank the list of articles according to a user's preference. Mathematically, the articles are ranked based on the probability of the user clicking on them. We model this as a binary classification problem which takes a user and an article as input and computes the probability of click i.e. 𝑃 𝑟𝑜𝑏(𝑐𝑙𝑖𝑐𝑘 = 1|𝑢𝑠𝑒𝑟, 𝑖𝑡𝑒𝑚). The input features for this model are a combination of the dense representation of the user and the article which we learn through graph embeddings.</p><p>During their activity on the application, users interact with many articles. More formally, the user-article interaction can be organized as bipartite graph 𝐺 = (𝑈, 𝐴, 𝐸), where 𝑈 denotes the set of users and 𝐴 denotes the set of articles. The set 𝑉 = 𝑈 ∪ 𝐴 is the set of nodes of 𝐺. There is an edge 𝑒 ∈ 𝐸 between a user 𝑢 ∈ 𝑈 and an article 𝑎 ∈ 𝐴 if the 𝑢 clicked on the snippet for 𝑎 to read the story. The set of all nodes connected to a node 𝑛, called the adjacency list of 𝑛 is denoted by 𝐸(𝑛).</p><p>Graph embedding for the bipartite graph: Graph representation learning techniques such as DeepWalk <ref type="bibr" target="#b4">[5]</ref> and Node2vec <ref type="bibr" target="#b2">[3]</ref> use random walks to embed a graph onto a low-dimensional space which maps each node to a dense vector. Following DeepWalk <ref type="bibr" target="#b4">[5]</ref>, we generate short truncated random walks starting from each node. The random walks are generated in a completely uniform and unbiased fashion, each adjacent node has equal probability to be picked. The random walks produce sequences of nodes of pre-decided length. For the bipartite graph, the random walks are generated by repeating the operations -1) Given the current node 𝑛 which is initialized at the starting node of the random walk, get its adjacency list 𝐸(𝑛). 2) Sample an edge from 𝐸(𝑛) which links Algorithm 1 Bipartite Graph Embedding Input: Bipartite graph 𝐺 = (𝑈, 𝐴, 𝐸), walk length 𝑙, walks per node 𝑟, embedding dimension 𝑑, context window size 𝑤.  Append curr walk to 𝑤𝑎𝑙𝑘𝑠 11: 𝑆𝑘𝑖𝑝𝐺𝑟𝑎𝑚(𝑤𝑎𝑙𝑘𝑠, 𝑤, 𝑑) Output: Embeddings of every node 𝑛 ∈ 𝑉 , 𝑣𝑛 ∈ R 𝑑 𝑛 to the node 𝑚. 3) Thee current node is updated to 𝑚 and the steps repeat. The procedure is described in Algorithm 1. The random walks on the bipartite graph have paths of the form</p><formula xml:id="formula_0">𝑈 𝑠𝑒𝑟 → 𝐴𝑟𝑡𝑖𝑐𝑙𝑒 → 𝑈 𝑠𝑒𝑟</formula><p>The random walks thus generated are sequences of the nodes, which can be thought as 'sentences' in an artificial language. The SkipGram <ref type="bibr" target="#b6">[7]</ref> model, developed for word embeddings, measures the probability of two words to co-occur within a fixed window on a sentence. It uses a fixed size window around every word to extract context and non-context words for the word under consideration. The model employs a single hidden layer neural network to learn a mapping from the word to its context word. In graph embeddings, the skip-gram model <ref type="bibr" target="#b6">[7]</ref> computes the probability of two nodes to co-occur within a fixed window on a random walk.</p><p>This procedure is an extension of the latent factor models <ref type="bibr" target="#b8">[10]</ref> of collaborative-filtering where matrix factorization is used to obtain dense representations of users and items in the same space. Graph embeddings have been shown to perform better than the matrix factorization for various tasks e.g. classification, clustering, link prediction etc. Mathematically, graph embeddings extend the contextual similarity from immediate neighborhood to truncated random walks. Graph embedding for the user-item bipartite graph, though effective, suffers from the problem of cold start i.e. it's unable to handle new content which was not the part of the graph. Content-based filtering has been an alternative to collaborativefiltering for handling the problem of fresh, unseen items. We merge collaborative and content filtering in a natural way in our setting by enriching the bipartite graph 𝐺 to a new graph 𝐺 ′ = (𝑈, 𝐴, 𝐶, 𝐸, 𝐸 ′ ) where 𝐶 denotes set of content nodes and 𝐸 ′ ) is the set of edges between the article and the content nodes. We use the named entities of the form place, person, organization as the content nodes. The named entities are at a finer level than topic modeling appraoch <ref type="bibr" target="#b18">[20]</ref> and are more constrained than the tf-idf based keyword filtering <ref type="bibr" target="#b19">[21]</ref>. The set of topics and keywords is not exhaustive, new articles can introduce topics and keywords, thus the models for extracting topics or keywords need to be retrained. Named Entity Recognition (NER), in turn, is article agnostic. It can extract entities in new articles without being retrained explicitly. In fact, there are off-shelf tools e.g. Spacy <ref type="bibr" target="#b20">[22]</ref> to extract names entities on any article. Thus, NER is cheaper to achieve than topic modeling and keyword extraction.</p><p>Graph embedding for the enriched graph: The enriched graph 𝐺 ′ contains user-user interaction via their shared interests in articles as well as the article-article interaction via the co-occurrence of entities in them. In the bipartite setting, the random walks transition between user and article nodes. While, in the enriched graph, the random walks are more complex. The random walker at the article node can either jump to the user node or to the entity node. The paths in the random walks can be of following types:</p><formula xml:id="formula_1">• 𝑈 𝑠𝑒𝑟 → 𝐴𝑟𝑡𝑖𝑐𝑙𝑒 → 𝑈 𝑠𝑒𝑟 • 𝑈 𝑠𝑒𝑟 → 𝐴𝑟𝑡𝑖𝑐𝑙𝑒 → 𝐸𝑛𝑡𝑖𝑡𝑦 → 𝐴𝑟𝑡𝑖𝑐𝑙𝑒 → 𝑈 𝑠𝑒𝑟 • 𝐴𝑟𝑡𝑖𝑐𝑙𝑒 → 𝑈 𝑠𝑒𝑟 → 𝐴𝑟𝑡𝑖𝑐𝑙𝑒 • 𝐴𝑟𝑡𝑖𝑐𝑙𝑒 → 𝐸𝑛𝑡𝑖𝑡𝑦 → 𝐴𝑟𝑡𝑖𝑐𝑙𝑒 → 𝑈 𝑠𝑒𝑟 • 𝐴𝑟𝑡𝑖𝑐𝑙𝑒 → 𝐸𝑛𝑡𝑖𝑡𝑦 → 𝐴𝑟𝑡𝑖𝑐𝑙𝑒 → 𝐸𝑛𝑡𝑖𝑡𝑦</formula><p>This provides a lot more possibilities then the 𝐴𝑟𝑡𝑖𝑐𝑙𝑒 → 𝑈 𝑠𝑒𝑟 and 𝑈 𝑠𝑒𝑟 → 𝐴𝑟𝑡𝑖𝑐𝑙𝑒 choices in the bipartite. Thus, the embeddings learned on these random walks are able to capture relationships between different types of nodes due to different types of ways they can be connected.</p><p>The enriched graph 𝐺 ′ is heterogeneous with two types of edges i.e. user-article and article-entity. The graph embedding methods like DeepWalk <ref type="bibr" target="#b4">[5]</ref> are restricted to homogeneous graphs, and may not be directly applicable to heterogeneous networks. One obvious problem is that by having unbiased random walks, we assign equal probability to each edge type. This increases the probability of the random walk going through an edge-type which has multiple edges from the current node. For any article node in the enriched graph, there are far more edges to the user nodes than to the entity nodes. Thus the random walk has higher chances of going to the user nodes, and in some cases completely avoid entity edges. This stems from an imbalance at the global level, we have a tens of thousands of entities, where as there are millions of users. We provide a simple way to resolve the problem of random walks being biased by the dominant edge type. The random walk is generated in two steps: (i) an edge type is chosen randomly from all possible edge types, (ii) an edge is randomly chosen from all edges of the selected edge type. This amounts to biasing the random walks uniformly with equal weights for each edge type. The procedure for learning representations of the nodes for the enriched graph is described in the Algorithm 2 and Algorithm 3.</p><p>In reality, different edge types contribute differently to the random walks and would have unequal weights. However, we do not have an intuitive way to obtain these weights. Some work has been done in this direction e.g. Metapath2vec metapath2vec, heterogeneous edge embeddings <ref type="bibr" target="#b21">[23]</ref> etc.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Algorithm 2 Enriched Graph Embedding</head><p>Input: Enriched graph 𝐺 = (𝑈, 𝐴, 𝐶, 𝐸, 𝐸 ′ ), walk length 𝑙, walks per node 𝑟, embedding dimension 𝑑, context window size 𝑤.  Generate a random number 𝑡 from 𝑈 𝑛𝑖𝑓 𝑜𝑟𝑚(0, 1).</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>7:</head><p>If 𝑡 &gt; 0.5 8:</p><formula xml:id="formula_2">next step = 𝑅𝑎𝑛𝑑𝑜𝑚𝑆𝑎𝑚𝑝𝑙𝑒(𝐸 ′ (curr node)) 9:</formula><p>if 𝑡 ≤ 0.5 10:</p><p>next step = 𝑅𝑎𝑛𝑑𝑜𝑚𝑆𝑎𝑚𝑝𝑙𝑒(𝐸(curr node)) Output: Next node in the random walk next step Recommendation Model: Having learned the embeddings for the users, items and the entities, we now describe the model for recommendation. We define the recommendation problem as a binary classification model which learns the probability of a user clicking on an article. We use logistic regression model on the embeddings learned on the graph. The input feature vector to the logistic regression model is the Hadamard product of the user feature vector and the article feature vector. If both the user and the article were present in the enriched graph i.e. we have an embedding for them, then the user and article feature vectors are their respective graph embeddings. The purpose of a news recommendation system is to recommend new, unseen articles, and we may not have an embedding for the fresh articles. In this case, we use the average of the embeddings of the entities present in the article as the article feature vector. Thus the input feature to the logistic regression is in the same space as the graph embedding space.</p><p>If 𝐹𝑢 ∈ R 𝑑 is the vector representation of user 𝑢 and 𝐹𝑎 ∈ R 𝑑 be the vector representation of article 𝑎, then the input to the logistic regression model is a vector 𝑥 (𝑢,𝑎) = (𝑥1, 𝑥2, . . . 𝑥 𝑑 ) ∈ R 𝑑 defined as:</p><formula xml:id="formula_3">(𝑥 (𝑢,𝑎) )𝑖 = (𝐹𝑢)𝑖 * (𝐹𝑎)𝑖<label>(1)</label></formula><p>The user feature vector is defined as the user embedding vector 𝑉𝑢 learned via graph embeddings. The article feature vector 𝐹𝑎 is given by the graph embedding 𝑉𝑎 ∈ R 𝑑 if 𝑎 has an embedding. Else, if 𝐶(𝑎) is the set of entities in the contents of 𝑎</p><formula xml:id="formula_4">𝐹𝑎 = 1 |𝐶(𝑎)| ∑︁ 𝑐∈𝐶(𝑎) 𝑉𝑐<label>(2)</label></formula><p>The pipeline for recommendation system described above as three components:</p><p>• Feature Learning: A large part of the user-article interaction data between time 𝑡0 and 𝑡1 is taken to build the enriched graph and then embeddings are learned on this graph. e.g. the user activity on the application between Jan 2016 to Jan 2019. An alternate interpretation of the recommendation model is as the link prediction <ref type="bibr" target="#b22">[24]</ref> in a user-article bipartite graph selected at a future time i.e. it contains users and new articles.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4">RESULTS AND ANALYSIS</head><p>In this section, we provide discussion on evaluation and analysis of the graph embeddings and the recommendation model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1">Experimental Setup</head><p>The data for this experiment was taken from a real-world news aggregation mobile application X 1 . The news content on X is presented as a scroll-able list which shows the headline and first couple of lines of the article. The user can read the full article by clicking on the shown snippet for that article.</p><p>During their activity on the platform, users click on the displayed article snippet if they want to read the full story or they scroll past it. We consider user clicking on an article as an indicator of their interest in the article and use the clicks as the positive data for the recommendation model. If a user scrolls past an article and does not click on it, we use that as the negative example. Thus, our model attempt to increase the click-through rate (CTR) as the metric. For learning the graph embeddings, we create an enriched graph using the activity of 500K users during the period of 3 months, adding 40K articles. There are also 6K entity nodes, which we extracted using the freely available Spacy API.</p><p>1 Name annonymized for the double-blind review. To learn embeddings of all the nodes in the enriched graph, we generate 30 random walks of length 100 for every node. The skip-gram model is trained using stochastic gradient (SGD) with a learning rate of 0.01 to minimize the negative sampling loss. Finally, we obtain an embedding of 128dimension for every node.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2">Multi-class Article Classification</head><p>We evaluate the feature representations obtained through graph embeddings on a standard supervised learning taskmulti-class classification of the news articles. News publishers often add some high level categories to the articles they publish e.g. Sports, Entertainment etc. We train a machine learning model on a the set of the labeled articles using their graph embeddings as input. The task is to predict the labels for the remaining articles. There are 8 news categories in our data -Business, Entertainment, Sports, Local, Tech, World, Lifestyle, Offbeat. In this experiment, we use a fraction of the labeled article nodes to train a multinomial logistic regression model with L2 regularization. Without explicitly using any content information, the performance of thee logistic regression model trained on node embeddings (node2vec) is similar to that of a model trained explicitly on content features e.g. word2vec <ref type="bibr" target="#b6">[7]</ref> or doc2vec <ref type="bibr" target="#b10">[12]</ref>. The comparison is presented in Table <ref type="table" target="#tab_1">1</ref>. The detailed results of the model are described in the confusion matrix in Table <ref type="table" target="#tab_2">2</ref>.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3">Recommendation Evaluation</head><p>The recommendation model is trained on a set of user-article pairs of positive examples (user clicked) and negative examples (no click) taken from the activity over a period of one month which is different from the data used for building the enriched graph. The logistic regression with L2 regularization is used for binary classification of the pairs as click or no-click. We then evaluate the model to predict whether a </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5">CONCLUSION AND FUTURE WORK</head><p>In this work, we proposed a graph-based news recommendation system which is a hybrid of collaborative-filtering and content information-filtering. Our method employs graph embeddings to automatically learn latent representation of users and articles. We extend the user-item bipartite graph to contain names entities from the articles. The entities are expected to capture the user preferences at a finer-level. The embedding methods bring user, items and the entities in the same space. We evaluated the learned latent representations via classification of article nodes into 8 high-level categories. We show that without explicitly using the contents of the article, we achieve results comparable to the NLP based features. We also design and evaluate a recommendation model as a binary classification model for computing the likelihood of the user clicking on the article. This model performs better than the hybrid collaborative-filtering and word embeddings based article similarity model. Though the model performs satisfactorily, this work has focused on simplicity since the goal of this paper is to show the efficacy of graph embedding methods for content recommendation. The embedding method has two hyperparameters -walk length and number of walks per node -which we chose based on heuristics and did not learn them. There is work using attention mechanism <ref type="bibr" target="#b25">[27]</ref> to learn the hyperparameters as an end-to-end system. The enriched network we used has two types of edges. We resolved the heterogeneity by giving equal importance to each edge type. Graph embeddings for heterogeneous networks is an active area of research e.g. metapath2vec <ref type="bibr" target="#b1">[2]</ref>, heterogeneous edge embeddings <ref type="bibr" target="#b21">[23]</ref> etc. The logistic regression model for recommendation was chosen its simplicity, a more suitable approach would be a neural network based model which is either trained on Siamese like loss (two input -user and item embeddings) <ref type="bibr" target="#b26">[28]</ref> or the triplet loss (three input -user prefers item 1 over item 2) <ref type="bibr" target="#b27">[29]</ref>. We plan to address some of these issues in a future work.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>1 : 4 :</head><label>14</label><figDesc>Initialize 𝑤𝑎𝑙𝑘𝑠 to 𝐸𝑚𝑝𝑡𝑦. 2: for 𝑖 = 0 to 𝑖 = 𝑟 do 3: for 𝑛 ∈ 𝑉 do Initialize curr walk to [𝑛].</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>5 : 7 : 9 :</head><label>579</label><figDesc>while 𝑠𝑡𝑒𝑝 = 𝑙 to 𝑙 do 6: curr step = curr walk[−1] adj list = 𝐸(curr step) 8: next step = 𝑅𝑎𝑛𝑑𝑜𝑚𝑆𝑎𝑚𝑝𝑙𝑒(adj list) Append next step to curr walk 10:</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head>1 : 4 :</head><label>14</label><figDesc>Initialize 𝑤𝑎𝑙𝑘𝑠 to 𝐸𝑚𝑝𝑡𝑦. 2: for 𝑖 = 0 to 𝑖 = 𝑟 do 3: for 𝑛 ∈ 𝑉 do Initialize curr walk to [𝑛].</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>5 : 7 : 9 :</head><label>579</label><figDesc>while 𝑠𝑡𝑒𝑝 = 𝑙 to 𝑙 do 6: curr step = curr walk[−1] next step = 𝑁 𝑒𝑥𝑡𝑁 𝑜𝑑𝑒(𝐺, curr step) 8: Append next step to curr walk Append curr walk to 𝑤𝑎𝑙𝑘𝑠 10: 𝑆𝑘𝑖𝑝𝐺𝑟𝑎𝑚(𝑤𝑎𝑙𝑘𝑠, 𝑤, 𝑑) Output: Embeddings of every node 𝑛 ∈ 𝑉 , 𝑣𝑛 ∈ R 𝑑 Algorithm 3 NextNode Input: Enriched graph 𝐺 = (𝑈, 𝐴, 𝐶, 𝐸, 𝐸 ′ ), curr node 1: If curr node ∈ 𝑈 2: next step = 𝑅𝑎𝑛𝑑𝑜𝑚𝑆𝑎𝑚𝑝𝑙𝑒(𝐸(curr node)) 3: else if curr node ∈ 𝐶 4: next step = 𝑅𝑎𝑛𝑑𝑜𝑚𝑆𝑎𝑚𝑝𝑙𝑒(𝐸 ′ (curr node)) 5: else if curr node ∈ 𝐴 6:</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head></head><label></label><figDesc>• Recommendation Model Training: The logistic regression model for the computing probability of a user clicking on an article is trained on a different sub-graph which comprises of interaction data between time 𝑡1 and 𝑡2 e.g. between Jan 2019 to Jun 2019. A portion of this training data is held out to validate the model. • Evaluation and Deployment: The trained model is then evaluated on the unseen data by considering the predictions of the model on usage between June 2019 to July 2019. The model is then deployed to the application.</figDesc><table /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 1 :</head><label>1</label><figDesc>Comparison of Embeddings</figDesc><table><row><cell cols="4">Features Precision Recall F-1</cell></row><row><cell>word2vec</cell><cell>0.90</cell><cell>0.92</cell><cell>0.91</cell></row><row><cell>doc2vec</cell><cell>0.91</cell><cell>0.92</cell><cell>0.91</cell></row><row><cell>node2vec</cell><cell>0.88</cell><cell>0.89</cell><cell>0.88</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 2 :</head><label>2</label><figDesc>Confusion Matrix of the Evaluation</figDesc><table><row><cell>Category</cell><cell cols="3">Precision Recall F-1</cell></row><row><cell>Business</cell><cell>0.83</cell><cell>0.71</cell><cell>0.77</cell></row><row><cell>Sports</cell><cell>0.90</cell><cell>0.92</cell><cell>0.91</cell></row><row><cell>Entertainment</cell><cell>0.97</cell><cell>0.95</cell><cell>0.96</cell></row><row><cell>Tech</cell><cell>0.86</cell><cell>0.83</cell><cell>0.84</cell></row><row><cell>Local</cell><cell>0.88</cell><cell>0.95</cell><cell>0.91</cell></row><row><cell>World</cell><cell>0.82</cell><cell>0.68</cell><cell>0.75</cell></row><row><cell>Lifestyle</cell><cell>0.93</cell><cell>0.85</cell><cell>0.89</cell></row><row><cell>Offbeat</cell><cell>0.87</cell><cell>0.62</cell><cell>0.72</cell></row><row><cell>Average</cell><cell>0.88</cell><cell>0.89</cell><cell>0.88</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_3"><head>Table 3 :</head><label>3</label><figDesc>Comparison of Recommendation Models -article pairs were clicked on not. We measure the performance of the model by computing the Area Under the ROC curve (AUC) and Precision at 5. The model shows a significant improvement over the previously used approach which is a combination of collaborative-filtering and word embeddings (Hybrid CF and Content). One major advantage is that our model is that it naturally blends collaborativefiltering and content-based filtering by putting both the user, item, and the content into the same space.</figDesc><table><row><cell>Model</cell><cell cols="2">AUC Precision@5</cell></row><row><cell>Hybrid CF and Content</cell><cell>0.72</cell><cell>0.89</cell></row><row><cell cols="2">Enriched graph embeddings 0.87</cell><cell>0.97</cell></row><row><cell>set of user</cell><cell></cell><cell></cell></row></table></figure>
		</body>
		<back>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">Heterogeneous network embedding via deep architectures</title>
		<author>
			<persName><forename type="first">S</forename><surname>Chang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Han</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Tang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">J</forename><surname>Qi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><forename type="middle">C</forename><surname>Aggarwal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><forename type="middle">S</forename><surname>Huang</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</title>
				<meeting>the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="119" to="128" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<analytic>
		<title level="a" type="main">Metapath2vec: Scalable representation learning for heterogeneous networks</title>
		<author>
			<persName><forename type="first">Y</forename><surname>Dong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">N</forename><forename type="middle">V</forename><surname>Chawla</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Swami</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</title>
				<meeting>the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="135" to="144" />
		</imprint>
	</monogr>
	<note>KDD &apos;17</note>
</biblStruct>

<biblStruct xml:id="b2">
	<monogr>
		<title level="m" type="main">node2vec: Scalable feature learning for networks</title>
		<author>
			<persName><forename type="first">A</forename><surname>Grover</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Leskovec</surname></persName>
		</author>
		<idno>CoRR abs/1607.00653</idno>
		<ptr target="http://arxiv.org/abs/1607.00653" />
		<imprint>
			<date type="published" when="2016">2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b3">
	<monogr>
		<title level="m" type="main">Representation learning on graphs: Methods and applications</title>
		<author>
			<persName><forename type="first">W</forename><forename type="middle">L</forename><surname>Hamilton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Ying</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Leskovec</surname></persName>
		</author>
		<idno>CoRR abs/1709.05584</idno>
		<ptr target="http://arxiv.org/abs/1709.05584" />
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Deepwalk: Online learning of social representations</title>
		<author>
			<persName><forename type="first">B</forename><surname>Perozzi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">R</forename><surname>Al-Rfou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Skiena</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</title>
				<meeting>the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>ACM</publisher>
			<date type="published" when="2014">2014</date>
			<biblScope unit="page" from="701" to="710" />
		</imprint>
	</monogr>
	<note>KDD &apos;14</note>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">A survey on network embedding</title>
		<author>
			<persName><forename type="first">P</forename><surname>Cui</surname></persName>
		</author>
		<author>
			<persName><forename type="first">X</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Pei</surname></persName>
		</author>
		<author>
			<persName><forename type="first">W</forename><surname>Zhu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">IEEE Transactions on Knowledge and Data Engineering</title>
		<imprint>
			<date type="published" when="2018">2018. 2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b6">
	<monogr>
		<title level="m" type="main">Efficient estimation of word representations in vector space</title>
		<author>
			<persName><forename type="first">T</forename><surname>Mikolov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">K</forename><surname>Chen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><surname>Corrado</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Dean</surname></persName>
		</author>
		<idno>CoRR abs/1301.3781</idno>
		<ptr target="http://arxiv.org/abs/1301.3781" />
		<imprint>
			<date type="published" when="2013">2013</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Google news personalization: scalable online collaborative filtering</title>
		<author>
			<persName><forename type="first">A</forename><forename type="middle">S</forename><surname>Das</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Datar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Garg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">S</forename><surname>Rajaram</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 16th international conference on World Wide Web</title>
				<meeting>the 16th international conference on World Wide Web</meeting>
		<imprint>
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<monogr>
		<title level="m" type="main">Matrix factorization for recommender systems</title>
		<author>
			<persName><forename type="first">Yehuda</forename><surname>Koren</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Robert</forename><surname>Bell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Chris</forename><surname>Volinsky</surname></persName>
		</author>
		<ptr target="https://datajobs.com/data-science-repo/Recommender-Systems-%5BNetflix%5D.pdf" />
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Group-Lens: Applying collaborative filtering to usenet news</title>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Konstan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">B</forename><forename type="middle">N</forename><surname>Miller</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Maltz</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">L</forename><surname>Herlocker</surname></persName>
		</author>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">R</forename><surname>Gordon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Riedl</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Commun. ACM</title>
		<imprint>
			<biblScope unit="volume">40</biblScope>
			<biblScope unit="page" from="77" to="87" />
			<date type="published" when="1997">1997</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Tomas Mikolov : Distributed Representations of Sentences and Documents</title>
		<author>
			<persName><forename type="first">V</forename><surname>Quoc</surname></persName>
		</author>
		<author>
			<persName><surname>Le</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 31 st International Conference on Machine Learning (ICML)</title>
				<meeting>the 31 st International Conference on Machine Learning (ICML)<address><addrLine>Beijing, China</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<monogr>
		<title level="m" type="main">User Modeling for Adaptive News Access, User Modeling and User-Adapted Interaction</title>
		<author>
			<persName><forename type="first">D</forename><surname>Billsus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">J</forename><surname>Pazzani</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2000">2000</date>
			<biblScope unit="volume">10</biblScope>
			<biblScope unit="page" from="147" to="180" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Learning User Profiles for Personalized Information Dissemination</title>
		<author>
			<persName><forename type="first">A</forename><surname>Tan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">C</forename><surname>Tee</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of 1998 IEEE International Joint conference on Neural Networks</title>
				<meeting>1998 IEEE International Joint conference on Neural Networks</meeting>
		<imprint>
			<date type="published" when="1998-05">May 1998</date>
			<biblScope unit="page" from="183" to="188" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">Evaluating adaptive user profiles for news classification</title>
		<author>
			<persName><forename type="first">R</forename><surname>Carreira</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">M</forename><surname>Crato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Gon?alves</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><forename type="middle">A</forename><surname>Jorge</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 9th international conference on Intelligent user interfaces</title>
				<meeting>the 9th international conference on Intelligent user interfaces</meeting>
		<imprint>
			<date type="published" when="2004">2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">A hybrid user model for news story classification</title>
		<author>
			<persName><forename type="first">D</forename><surname>Billsus</surname></persName>
		</author>
		<author>
			<persName><forename type="first">;</forename><forename type="middle">M</forename><surname>Pazzani</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the Seventh International Conference on User Modeling</title>
				<meeting>the Seventh International Conference on User Modeling</meeting>
		<imprint>
			<date type="published" when="1999">1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<analytic>
		<title level="a" type="main">Combining Content-Based and Collaborative Filters in an Online Newspaper</title>
		<author>
			<persName><forename type="first">M</forename><surname>Claypool</surname></persName>
		</author>
		<author>
			<persName><forename type="first">A</forename><surname>Gokhale</surname></persName>
		</author>
		<author>
			<persName><forename type="first">T</forename><surname>Miranda</surname></persName>
		</author>
		<author>
			<persName><forename type="first">P</forename><surname>Murnikov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">D</forename><surname>Netes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">M</forename><surname>Sartin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of ACM SIGIR Workshop on Recommender Systems</title>
				<meeting>ACM SIGIR Workshop on Recommender Systems</meeting>
		<imprint>
			<date type="published" when="1999">1999</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">Pixie: A System for Recommending 3+ Billion Items to 200+ Million Users in Real-Time</title>
		<author>
			<persName><forename type="first">Chantat</forename><surname>Eksombatchai</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Pranav</forename><surname>Jindal</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jerry</forename><forename type="middle">Zitao</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yuchen</forename><surname>Liu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rahul</forename><surname>Sharma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Charles</forename><surname>Sugnet</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mark</forename><surname>Ulrich</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jure</forename><surname>Leskovec</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Inductive Representation Learning on Large Graphs</title>
		<author>
			<persName><forename type="first">William</forename><forename type="middle">L</forename><surname>Hamilton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rex</forename><surname>Ying</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jure</forename><surname>Leskovec</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">31st Conference on Neural Information Processing Systems (NIPS 2017)</title>
				<meeting><address><addrLine>Long Beach, CA, USA</addrLine></address></meeting>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<monogr>
		<title level="m" type="main">Collaborative Topic Modeling for Recommending Scientific Articles, KDD11</title>
		<author>
			<persName><forename type="first">Chong</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Chong</forename><surname>Wang</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2011">August 2124, 2011</date>
			<pubPlace>San Diego, California, USA</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Selecting keywords for content based recommendation</title>
		<author>
			<persName><forename type="first">Christian</forename><surname>Wartena</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Wout</forename><surname>Slakhorst</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Martin</forename><surname>Wibbels</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 19th ACM international conference on Information and knowledge management (CIKM)</title>
				<meeting>the 19th ACM international conference on Information and knowledge management (CIKM)<address><addrLine>Toronto, ON, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2010">October 26 -30, 2010</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<monogr>
		<ptr target="https://spacy.io/" />
		<title level="m">Spacy -Industrial-Strength Natural Language Processing</title>
				<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<monogr>
		<title level="m" type="main">Heterogeneous Edge Embeddings for Friend Recommendation</title>
		<author>
			<persName><forename type="first">Janu</forename><surname>Verma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Srishti</forename><surname>Gupta</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Debdoot</forename><surname>Mukherjee</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tanmoy</forename><surname>Chakraborty</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2019-04">April 2019</date>
			<pubPlace>ECIR Cologne</pubPlace>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">The link-prediction problem for social networks</title>
		<author>
			<persName><forename type="first">D</forename><surname>Liben-Nowell</surname></persName>
		</author>
		<author>
			<persName><forename type="first">J</forename><surname>Kleinberg</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of the American society for information science and technology</title>
		<imprint>
			<biblScope unit="volume">58</biblScope>
			<biblScope unit="issue">7</biblScope>
			<biblScope unit="page" from="1019" to="1031" />
			<date type="published" when="2007">2007</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">hdbscan: Hierarchical density based clustering</title>
		<author>
			<persName><forename type="first">Leland</forename><surname>Mcinnes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">John</forename><surname>Healy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Steve</forename><surname>Astels</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">The Journal of Open Source Software</title>
		<imprint>
			<biblScope unit="volume">2</biblScope>
			<biblScope unit="issue">11</biblScope>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Visualizing High-Dimensional Data Using t-SNE</title>
		<author>
			<persName><forename type="first">L</forename><forename type="middle">J P</forename><surname>Van Der Maaten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">G</forename><forename type="middle">E</forename><surname>Hinton</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="2579" to="2605" />
			<date type="published" when="2008-11">Nov. 2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Watch Your Step: Learning Node Embeddings via Graph Atten</title>
		<author>
			<persName><forename type="first">Sami</forename><surname>Abu-El-Haija</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Bryan</forename><surname>Perozzi</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rami</forename><surname>Al-Rfou</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alex</forename><surname>Alemi</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">tion, 32nd Conference on Neural Information Processing Systems (NeurIPS 2018)</title>
				<meeting><address><addrLine>Montral, Canada</addrLine></address></meeting>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">DeepFace: Closing the Gap to Human-Level Performance in Face Verification</title>
		<author>
			<persName><forename type="first">Yaniv</forename><surname>Taigman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ming</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Marc</forename><forename type="middle">'</forename></persName>
		</author>
		<author>
			<persName><forename type="first">Aurelio</forename><surname>Ranzato</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lior</forename><surname>Wolf</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition</title>
				<meeting>the IEEE Computer Society Conference on Computer Vision and Pattern Recognition</meeting>
		<imprint>
			<date type="published" when="2014">2014</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">FaceNet: A Unified Embedding for Face Recognition and Clustering</title>
		<author>
			<persName><forename type="first">Florian</forename><surname>Schroff</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Dmitry</forename><surname>Kalenichenko</surname></persName>
		</author>
		<author>
			<persName><forename type="first">James</forename><surname>Philbin</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition</title>
				<meeting>the IEEE Computer Society Conference on Computer Vision and Pattern Recognition</meeting>
		<imprint>
			<date type="published" when="2015">2015</date>
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
