<?xml version="1.0" encoding="UTF-8"?>
<TEI xml:space="preserve" xmlns="http://www.tei-c.org/ns/1.0" 
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" 
xsi:schemaLocation="http://www.tei-c.org/ns/1.0 https://raw.githubusercontent.com/kermitt2/grobid/master/grobid-home/schemas/xsd/Grobid.xsd"
 xmlns:xlink="http://www.w3.org/1999/xlink">
	<teiHeader xml:lang="en">
		<fileDesc>
			<titleStmt>
				<title level="a" type="main">Citation Intent Classification Through Weakly Supervised Knowledge Graphs</title>
			</titleStmt>
			<publicationStmt>
				<publisher/>
				<availability status="unknown"><licence/></availability>
			</publicationStmt>
			<sourceDesc>
				<biblStruct>
					<analytic>
						<author>
							<persName><forename type="first">Xinwei</forename><surname>Du</surname></persName>
							<email>xinweidu@usc.edu</email>
							<affiliation key="aff0">
								<orgName type="department">Information Sciences Institute</orgName>
								<address>
									<addrLine>Marina del Ray</addrLine>
									<region>CA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">University of Southern California</orgName>
								<address>
									<settlement>Los Angeles</settlement>
									<region>CA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Kian</forename><surname>Ahrabian</surname></persName>
							<email>ahrabian@usc.edu</email>
							<affiliation key="aff0">
								<orgName type="department">Information Sciences Institute</orgName>
								<address>
									<addrLine>Marina del Ray</addrLine>
									<region>CA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">University of Southern California</orgName>
								<address>
									<settlement>Los Angeles</settlement>
									<region>CA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Arun</forename><surname>Baalaaji</surname></persName>
						</author>
						<author>
							<persName><forename type="first">Sankar</forename><surname>Ananthan</surname></persName>
							<affiliation key="aff0">
								<orgName type="department">Information Sciences Institute</orgName>
								<address>
									<addrLine>Marina del Ray</addrLine>
									<region>CA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">University of Southern California</orgName>
								<address>
									<settlement>Los Angeles</settlement>
									<region>CA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Richard</forename><forename type="middle">Delwin</forename><surname>Myloth</surname></persName>
							<email>myloth@usc.edu</email>
							<affiliation key="aff0">
								<orgName type="department">Information Sciences Institute</orgName>
								<address>
									<addrLine>Marina del Ray</addrLine>
									<region>CA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">University of Southern California</orgName>
								<address>
									<settlement>Los Angeles</settlement>
									<region>CA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<persName><forename type="first">Jay</forename><surname>Pujara</surname></persName>
							<email>jpujara@usc.edu</email>
							<affiliation key="aff0">
								<orgName type="department">Information Sciences Institute</orgName>
								<address>
									<addrLine>Marina del Ray</addrLine>
									<region>CA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
							<affiliation key="aff1">
								<orgName type="institution">University of Southern California</orgName>
								<address>
									<settlement>Los Angeles</settlement>
									<region>CA</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<author>
							<affiliation key="aff2">
								<orgName type="laboratory">The Third AAAI Workshop on Scientific Document Understanding 2023</orgName>
								<address>
									<addrLine>February 14th</addrLine>
									<postCode>2023</postCode>
									<settlement>Washington</settlement>
									<region>DC</region>
									<country key="US">USA</country>
								</address>
							</affiliation>
						</author>
						<title level="a" type="main">Citation Intent Classification Through Weakly Supervised Knowledge Graphs</title>
					</analytic>
					<monogr>
						<idno type="ISSN">1613-0073</idno>
					</monogr>
					<idno type="MD5">24F756E5A22E9A7D1B4D4382D5FB6C31</idno>
				</biblStruct>
			</sourceDesc>
		</fileDesc>
		<encodingDesc>
			<appInfo>
				<application version="0.7.2" ident="GROBID" when="2025-04-23T16:26+0000">
					<desc>GROBID - A machine learning software for extracting information from scholarly documents</desc>
					<ref target="https://github.com/kermitt2/grobid"/>
				</application>
			</appInfo>
		</encodingDesc>
		<profileDesc>
			<textClass>
				<keywords>
					<term>Citation Intent Classification</term>
					<term>Knowledge Graphs</term>
					<term>Graph Neural Networks</term>
					<term>Large Language Models</term>
					<term>Weakly supervised learning</term>
				</keywords>
			</textClass>
			<abstract>
<div xmlns="http://www.tei-c.org/ns/1.0"><p>Citations are scientists' tools for grounding their innovations and findings in the existing collective knowledge. They are used for semantically distinct purposes as scientists utilize them at different parts of their work to convey specific information. As a result, a crucial aspect of scientific document understanding is recognizing the authorial intent associated with citations. Current state-of-the-art methods rely on contextual sentences surrounding each citation to classify the intent. However, in the absence of textual content, these approaches become unusable. In this work, we propose a text-free citation intent classification method built on relational information among scholarly works in this work. To this end, we introduce a large-scale knowledge graph built from the publications in the SciCite dataset and their multi-hop neighborhood extracted from The Semantic Scholar Open Research Corpus (S2ORC). We also augment this knowledge graph by adding weakly-labeled links based on the intent information available in the S2ORC. Finally, we cast the intent classification task as a link prediction problem on the newly created knowledge graph. We study this problem in both transductive and inductive settings. Our experimental results show that we can achieve a comparable macro F1 score to word embedding content-based methods by only relying on features and relations derived from this knowledge graph. Specifically, we achieve macro F1 scores of 62.16 and 59.81 in the transductive and inductive settings, respectively, on the link-level SciCite dataset. Moreover, by combining our method with the state-of-the-art NLP-based model, we achieve improvements across all metrics.</p></div>
			</abstract>
		</profileDesc>
	</teiHeader>
	<text xml:lang="en">
		<body>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="1.">Introduction</head><p>Citations are the primary way of identifying past contributions and connecting progress in new publications to existing literature. Nevertheless, not all citations indicate the same meaning. Authors use citations sparingly with specific intent behind them. For example, some papers are cited for providing background information in a domain, while others are cited when adopting or adapting a previously-used methodology. There are also scenarios where the same paper is used as background information and methodology use-case in different contexts simultaneously. Understanding citation intent is crucial to studying scholarly works, given the universality of using citations. Current state-of-the-art citation intent classification models <ref type="bibr" target="#b17">[17,</ref><ref type="bibr" target="#b0">1,</ref><ref type="bibr" target="#b4">4]</ref> rely heavily on textual information, e.g., the sentences surrounding the citation. However, such information is expensive to obtain and in some scenarios inaccessible altogether. Consequently, we need models that could operate without having access to textual information. Previous works <ref type="bibr" target="#b3">[3,</ref><ref type="bibr" target="#b26">26,</ref><ref type="bibr" target="#b6">6]</ref> have shown the importance of relational and structural information available in links among publications for various tasks. In this work, we propose a general citation intent classification method that relies purely on structural information.</p><p>Besides helping researchers better understand the relationship among publications, citation intent analysis has been used for studying various other aspects of scientific works such as research domain evolution <ref type="bibr" target="#b10">[10]</ref>, scientific impact analysis <ref type="bibr" target="#b19">[19]</ref>, scientific document summarization <ref type="bibr" target="#b5">[5]</ref>, and retrieving related scientific works <ref type="bibr" target="#b16">[16]</ref>. The main three categories of citations are "Result, " "Method, " and "Background" <ref type="bibr" target="#b4">[4]</ref>. These categories describe the reasons behind making a scientific connection, referencing a publication in another publication. Classifying citations into these groups has traditionally required a high level of expertise in the respective scientific domains. This constraint, combined with the high cost of expert human labor, has resulted in highly scarce datasets, which makes the task even more difficult.</p><p>Previous works have proposed classifying citation intent through feature engineering-based <ref type="bibr" target="#b10">[10]</ref> and representation learning-based <ref type="bibr" target="#b0">[1]</ref> methods. However, most of these methods depend on textual information. As a result, they require a complex multi-stage pipeline of parsing documents, identifying citation contexts, and predicting citation intent <ref type="bibr" target="#b13">[13]</ref>. Besides being prone to error propagation from various pipeline stages, the use of these models is limited to situations where the full text is available in a proper format. This work introduces a pure graph-based approach to classifying citation intent. We extend the existing SciCite dataset with 2-hop neighborhoods extracted from The Semantic Scholar Open Research Corpus (S2ORC). To further enrich the graph, we utilize the intent information provided in the S2ORC to create a weakly supervised knowledge graph (KG) consisting of the publications and the relations that match the provided intents. Our main idea is to use contextualized relational patterns to make predictions, obviating the need for textual context. Given the newly built KG, we cast the intent classification problem into the common link prediction problem on KGs. Specifically, we train a model to learn representations for entities and relations. Using these representations, we run the following query on the KG: (𝑠, ?, 𝑜), where 𝑠 cites 𝑜.</p><p>Converting this problem into a link prediction task allows us to adapt and extend widely used KG embedding models to this problem. We study the link prediction problem in both transductive and inductive settings. Our experimental results show that although our KGbased method underperforms compared to the large language model-based approaches, it is comparable or even superior to the word embedding-based methods. Moreover, our experiments with combining the NLP-based and graph-based methods show slight improvements over the current state-of-the-art model. These findings further signify the importance of relational patterns for citation intent classification.</p><p>The contributions of this work are as follows:</p><p>1. Extending the SciCite dataset using the S2ORC dataset to generate a large-scale weakly supervised KG. 2. Introducing a novel graph-based approach for citation intent classification built on top of the newly built KG. 3. Presenting benchmarks for both transductive and inductive settings. 4. Presenting analyses on the effect of different parts of the methodology such as weak supervision and feature engineering.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.">Related Work</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.1.">Citation Function/Intent Schemes</head><p>Many prior works have studied the problem of creating categorical schemes for citation intent which in some works is referred to as citation function <ref type="bibr" target="#b9">[9]</ref>. Earlier works were focused on creating more fine-grained categories, going as far as defining 35 <ref type="bibr" target="#b7">[7]</ref> and 12 <ref type="bibr" target="#b21">[21]</ref> fine-grained schemes for scientific arguments. The more recent works however have focused on creating more concise categories. For example, ACL-ARC <ref type="bibr" target="#b10">[10]</ref> proposes a 6-class intent categorization scheme: Background, Motivation, Uses, Extension, Comparison or Contrast, and Future. SciCite <ref type="bibr" target="#b4">[4]</ref> is even more restrictive and drops or combines small fine-grained classes to provide a more concise 3-class annotation scheme: Background, Method, and Result.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.2.">Citation Intent Classification Methods</head><p>Before the explosion of deep learning approaches, most methods relied on a combination of hand-crafted features and classic machine learning models. For example, in one instance <ref type="bibr" target="#b23">[23]</ref>, authors propose 12 different features, including citation count, PageRank value, and author overlap, and use classic machine learning models such as SVM and Random Forest for classification. In another instance <ref type="bibr" target="#b10">[10]</ref>, authors define pattern-based, topic-based, and prototypical argument features and use SVM to make predictions.</p><p>With the advent of deep learning models and the emergence of large language models in recent years, representation learning-based methods have outperformed the hand-crafted methods achieving a higher accuracy by considering the textual information. Recent works have proposed the use of structural scaffolds <ref type="bibr" target="#b4">[4]</ref>, BERT-based models trained on the scientific corpus (SciBERT) <ref type="bibr" target="#b0">[1]</ref>, word embedding-based approaches <ref type="bibr" target="#b17">[17]</ref>, and creating a heterogeneous context graph based on an academic </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="2.3.">Knowledge Graph Embedding Models</head><p>KGs are structured information repositories consisting of a set of nodes representing entities and a set of typed edges representing relations. Since, in most cases, the KG nodes and edges are not attributed, KG embedding (KGE) models aim to learn low-dimensional representations for all entities and relations. The most common traditional shallow KGE methods are TransE <ref type="bibr" target="#b1">[2]</ref>, Com-plEx <ref type="bibr" target="#b22">[22]</ref>, and RotatE <ref type="bibr" target="#b20">[20]</ref>. More recent GNN-based KGE methods leverage the message-passing scheme of GNNs, enabling more complex multi-hop reasoning. Examples of these methods are GCN <ref type="bibr" target="#b11">[11]</ref>, which leverages the spectral information for information propagation but is limited to mono-relational KGs, R-GCN <ref type="bibr" target="#b18">[18]</ref>, which extends GCN to support multi-relational KGs, and Graph-SAGE <ref type="bibr" target="#b8">[8]</ref> which introduces an inductive framework to handle unseen nodes.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.">Dataset</head><p>The SciCite dataset focuses on individual citation links and ignores the significance of broader relational connections and features. To overcome this issue, we construct a knowledge graph by mapping each entity in the SciCite dataset to the S2ORC and adding their 2-hop citation neighborhoods. The S2ROC contains more than 206 million publications and 2.49 billion citation links. Apart from the regular citation links, this corpus provides partial intent labels for citations using a 3-class scheme as follows:</p><p>1. Background: Describe a problem, topic, or concept 2. Method: Provide a method, tool, or dataset 3. Result: To make a comparison Moreover, the SciCite dataset is tailored for sentence classification methods, where input features are textual excerpts and the output labels are citation intents. We reformulate this task as link prediction on KGs, where the input features are a representation of the source (citing) paper and the target (cited) paper, and the output is the label of a citation link between the source and target. We release all our datasets under a CC-BY-SA license at TBD</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.1.">Entity Mapping</head><p>We first map each paper in the SciCite dataset to the S2ORC by matching SciCite's IDs to Semantic Scholar's SHA IDs. Since a publication could have many SHA IDs and only one Corpus ID, we then map each SHA ID to the unique Corpus ID to extract unique entities. From the 13,080 papers with unique IDs in SciCite, we successfully map 13,019 of them to valid SHA IDs in semantic scholar, while the remaining 61 papers do not have any corresponding records. We believe this is due to publication removals, as the SciCite dataset was created from the S2ORC in 2019. After converting SHA IDs to Corpus IDs, we end up with 13,011 unique entities and 8 duplicate entities.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="3.2.">Dataset Splitting</head><p>The original SciCite dataset contains 11,020 humanlabeled samples. Hence, to adapt it to our link prediction setting, we reconstruct two datasets: SciCiteorigin and SciCite resplit . SciCiteorigin adheres to the same benchmarks reported in prior works but is modified to remove overlapping citation links in the training and test sets. To maximize usage of the training data while removing artifacts, we create SciCite resplit that performs additional cleaning, provides a stronger separation of training and test sets, and avoids multi-intent citations. Table <ref type="table" target="#tab_0">1</ref> showcases the statistic of these datasets.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>SciCiteorigin:</head><p>To make methods comparable, we use the same validation and test sets as SciCite for this dataset and try to keep the training set as close as possible. We convert each publication in the SciCite dataset to a Semantic Scholar entity using the mapped Corpus IDs and drop the contextual sentence-level information. We assign a random unique ID to publications without a Corpus ID. After this procedure, we end up with a set of links for our link prediction task.</p><p>Due to the removal of the contextual information, some of the training links appear exactly the same in the test set. Hence, we remove 641 training set samples that also appear in the test set to prevent data leakage. Moreover, since only one link in the test set has multiple intents, we treat the link prediction problem as a multi-class task rather than a multi-label task. In this scenario, the multi-intent links are represented as separate samples with the same inputs and different outputs. Multi-label methods may be a promising future extension of our work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>SciCite resplit :</head><p>Even though we convert the SciCite dataset to the SciCiteorigin, problems, such as duplicate citations and multi-label links, still exist. Therefore, we further tailor the SciCite dataset to create a better link prediction dataset for graph-based models. First, we remove all the entities, and their related samples, that do not have a mapped Corpus ID. Then, similar to SciCiteorigin, we convert the remaining samples to a set of links. Following this, we drop all duplicate samples. Among the remaining 6,458 unique links, 5,886 only have one intent, 489 have two intents, and 83 have all three intents. We remove all the multi-intent links and resplit the dataset with ratios of 70%/15%/15% for training, validation, and test sets, respectively.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.">Method</head><p>Throughout the rest of this work, for simplicity, we use the term publication to denote all types of academic publications, e.g., books and papers. Moreover, we use the terms citation and reference to denote incoming and outgoing links, respectively.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.1.">Weak Supervision</head><p>In order to enrich our data and provide more information to the models, we extract the set of intents provided in the S2ORC dataset for each citation link. The intent labels in S2ORC are extracted using the structural scaffolds model <ref type="bibr" target="#b4">[4]</ref> at a sentence level. In this scenario, we implicitly use the existing data derived from the content for bootstrapping our approach. We refer to these links as weakly labeled due to being labeled by a noisy model rather than a human expert. Since the intent labels are partial at a sentence level, citation links could have zero intent in the absence of text or several intents in an abundance of use cases.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.2.">Knowledge Graph Construction</head><p>Given the S2ORC dataset, we expand the SciCite dataset using the mapped entities to construct a KG containing 2-hop neighborhoods of the publications. Figure <ref type="figure" target="#fig_0">1</ref> illustrates an overview of the expanded KG. This work uses the 2022-09-13 version of the corpus downloaded from the bulk API. Formally, given the set of mapped entities 𝒱0, the set of 𝑘-hop nodes 𝒱 𝑘 is defined as</p><formula xml:id="formula_0">𝒱 𝑘 = 𝒱 𝑘−1 ∪ {𝑦 | ∃𝑥 ∈ 𝒱 𝑘−1 : 𝑦 ∈ 𝒩𝑥}<label>(1)</label></formula><p>where for a given entity 𝑥, 𝒩𝑥 denotes all the entities that cite or are cited by 𝑥, i.e., the set of neighboring entities. Given the sets of unlabeled links 𝒰 and weakly labeled links ℒ, the set of 𝑘-hop edges ℰ 𝑘 is defined as</p><formula xml:id="formula_1">ℰ 𝒰 𝑘 = {(𝑥, 𝑦, UNK) | 𝑥, 𝑦 ∈ 𝒱 𝑘 , (𝑥, 𝑦) ∈ 𝒰} (2) ℰ ℒ 𝑘 = ∪𝑟{(𝑥, 𝑦, 𝑟) | 𝑥, 𝑦 ∈ 𝒱 𝑘 , (𝑥, 𝑦) ∈ ℒ𝑟} (3) ℰ 𝑘 = ℰ 𝒰 𝑘 ∪ ℰ ℒ 𝑘<label>(4)</label></formula><p>where 𝑟 ∈ {Background, Method, Result} and ℒ𝑟 denotes the set of all weakly labeled links with label 𝑟. Consequently, given the sets of 𝑘-hop nodes 𝒱 𝑘 and edges ℰ 𝑘 , the extracted 𝑘-hop KG, 𝒢 𝑘 , is defined as</p><formula xml:id="formula_2">𝒢 𝑘 = (𝒱 𝑘 , ℰ 𝑘 )<label>(5)</label></formula><p>The specific statistics of the extracted KG and the original semantic scholar corpus are reported in Table <ref type="table" target="#tab_1">2</ref>. Since not every link has weakly labeled intent, this table also provides the percentage of weakly labeled links for each corresponding graph. Although we extract 𝒢2, given its scale, we opt to run our current experiment only on 𝒢1 and leave the larger-scale experiments for future works.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.3.">Feature Engineering</head><p>Since none of the publications in our KGs have any features or pre-defined representation, we propose to represent them through their references, citations, and graphbased features. More specifically, from S2ROC we extract the in-degrees and out-degrees of citations (or references), background links, method links, and result links. As a result, each paper is represented with an 8-dimensional feature vector, 4 for each in-degree and out-degree feature. For the publications where the content is unavailable, the out-degree intent-based features will be zero since those features are based on the noisy sentence-level model that the Semantic Scholar uses. However, the in-degree features may not be zero as long as the citing paper's content is available. For the new publications, i.e., unseen nodes in the inductive setting, the only known non-zero feature is the reference count. We normalize the reference and citation features by a biased log factor defined as</p><formula xml:id="formula_3">ℎ ¯𝑥 = log 10 (ℎ𝑥 + 1 + 𝛼)<label>(6)</label></formula><p>where 𝛼 is a bias hyperparameter. We specifically set 𝛼 = −0.9 to get a normalized value of −1 for zeroreference and zero-citation situations. Moreover, we normalize the non-zero in-degree intentbased features into a [0, 1] probability distribution as follows:</p><formula xml:id="formula_4">ℎ ¯𝑥 = ℎ𝑥 ℎ Background + ℎ Method + ℎ Result (7)</formula><p>The same normalization step is used for out-degree features separately.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.4.">Baselines Knowledge Graph Embedding Models:</head><p>Traditional KGE models consist of two shallow embeddings as entity and relation encoders and a score function as a decoder to predict the likelihood of a link. These models are trained in a contrastive way by masking either one of the entities in a given triplet (head, relation, tail) and sampling a set of negative entities, contrasting the positive entity.</p><p>Since the traditional KGE methods rely on shallow embeddings for encoding entities and relations, they can only be used in the transductive setting and cannot operate on unseen nodes. For our experiments, we use the available implementations of TransE, ComplEx, and Ro-tatE in the DGL-KE toolkit <ref type="bibr" target="#b27">[27]</ref>. In the evaluation phase, we calculate the likelihood of all different relation types for each link and consider the highest likelihood as the model's intent prediction.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Hybrid Models:</head><p>To increase the reasoning power of the traditional KGE models, we devise a two-stage approach based on multilayer perceptron (MLP). We first use the traditional KGE models to learn embeddings for entities and relations. Then, instead of relying on the produced likelihood scores, we concatenate the vectors of two entities and pass that through an MLP to get logit values. Formally, given a link (𝑢, 𝑣) and their respective learned representation (𝑧𝑢, 𝑧𝑣), we calculate the logit values as</p><formula xml:id="formula_5">𝑝 = MLP([𝑧𝑢‖𝑧𝑣])<label>(8)</label></formula><p>where 𝑝 ∈ R 𝒞 contains the unnormalized logits for each class. The predicted class 𝑐 is then calculated as argmax 𝑐 sigmoid(𝑝). ( <ref type="formula">9</ref>)</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Natural Language Processing Models:</head><p>We include the reported results of several state-of-the-art Natural Language Processing (NLP) methods. Specifically, we include results from the word embedding-based methods such as Infersent-KMeans, Infersent-HDBSCAN, Glove-KMeans, and Glove-HDBSCAN <ref type="bibr" target="#b17">[17]</ref>, BiLSTMbased method Structural Scaffolds <ref type="bibr" target="#b4">[4]</ref>, and large language model-based method SciBERT <ref type="bibr" target="#b0">[1]</ref>. Moreover, we report the results of fine-tuning a pre-trained SciBERT model on both datasets. All these methods use textural information and are evaluated on the SciCite dataset.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="4.5.">Multi-Hop Link Prediction (MHLP)</head><p>Transductive and inductive settings are the most common link prediction evaluating schemes for KGs. The main difference between these two settings is having a fixed set of nodes in both the training and evaluation phases (transductive) versus allowing the addition of unseen nodes in the evaluation phase (inductive). This work refers to citation intent prediction on unseen publications as the inductive setting, whereas the transductive setting refers to citation intent prediction on already seen publications.</p><p>We propose an adaptable graph-based model for citation intent prediction in both the transductive and inductive settings. The primary basis of this approach is that a node, i.e., publication, could be represented as a combination of the neighboring nodes' representations. Let ℎ (0) 𝑥 be the extracted feature vector for any arbitrary node 𝑥. We calculate the representation of an arbitrary node 𝑣 at layer 𝑙 + 1 of a multilayer model as</p><formula xml:id="formula_6">ℎ (𝑙+1) 𝒩𝑣 = 1 |𝒩𝑣| ∑︁ 𝑢∈𝒩𝑣 ℎ (𝑙) 𝑢 (10) ℎ (𝑙+1) 𝑣 = 𝜎(𝑊 (𝑙+1) [ℎ (𝑙) 𝑣 ‖ℎ (𝑙+1) 𝒩𝑣 ])<label>(11)</label></formula><p>where 𝜎 is a non-linear function. Throughout our experiments, we specifically use ReLU to introduce nonlinearity. Given the node representation from a 𝐿-layer model and a link (𝑢, 𝑣), we calculate the logit values as</p><formula xml:id="formula_7">𝑝 = MLP([ℎ (𝐿) 𝑢 ‖ℎ (𝐿) 𝑣 ])<label>(12)</label></formula><p>where 𝑝 ∈ R 𝒞 contains the unnormalized logits for each class and 𝒞 is the set of all classes. The predicted class 𝑐 is then calculated as</p><formula xml:id="formula_8">argmax 𝑐 sigmoid(𝑝). (<label>13</label></formula><formula xml:id="formula_9">)</formula><p>The main disadvantage of the inductive settings is that the unseen nodes only have one available feature, i.e., reference count. This absence of information makes the task extremely difficult, as the feature vectors are highly sparse. However, our model tries to diminish this effect by using the message-passing scheme, as defined in Equation 11, to aggregate information through connected entities, i.e., cited papers, creating a denser representation for the unseen nodes.</p><p>All our models are trained using the cross-entropy loss defined as</p><formula xml:id="formula_10">𝑙𝑛 = − log exp(𝑝𝑦 𝑛 ) ∑︀ |𝒞| 𝑖=1 exp(𝑝𝑖)<label>(14)</label></formula><p>where and 𝑝𝑥 is the logit value for class 𝑥 given the prediction vector 𝑝.  </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Composite Model:</head><p>To further test the capabilities of our proposed model and use both structural and textual information, we devise a multi-modal model comprising encoders for both the graph structure and the citation context. Specifically, we use a pre-trained SciBERT model for encoding the citation phrase text and our MHLP model for encoding the citation graph around the citation link. Figure <ref type="figure" target="#fig_1">2</ref> illustrates an overview of the composite model.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.">Experiments</head><p>In this section, we report our experimental results on both of the SciCiteorigin and SciCite resplit datasets. All the graphbased experiments are carried out on the 𝒢1 KG. For the traditional KGE methods, we tune their hyperparameters as described in Appendix A.1 and train them using the hyperparameters showcased in Table <ref type="table">4</ref>. For the hybrid methods, the KGE component is first trained to generate node features using the hyperparameters described in Table <ref type="table">4</ref>. Then, the MLP component is trained using the procedure described in A.2 to predict the citation intent. For the MHLP-based methods, in both transductive and inductive settings, we use a 1-layer variation on top of the normalized features extracted as described in Section 4.3. Moreover, we tune their hyperparameters and train them as described in Appendix A.3. For the SciBERT method, we freeze the pre-trained model and add an MLP module on top of the 768-dimensional [CLS] token output. Similar to the other models, the MLP module is tuned using the parameters described in A.2. For the composite model, during the training phase, we freeze the SciBERT model in the first two epochs as a warm-up step for the graph encoder; then, we jointly train both encoders along with the final prediction module.</p><p>To control for the effect of the pre-training using traditional KGE models, we also run a variation with randomly initialized node features and designate it as "Random + MLP." For the NLP models, we use the previously reported results <ref type="bibr" target="#b17">[17]</ref> to compare our models on the test set-aligned SciCiteorigin dataset. Finally, we also include the results from random and most common class predictions as sanity checks. All the models are implemented using PyTorch <ref type="bibr" target="#b14">[14]</ref> and trained on a machine with a single Quadro RTX 8000 GPU, 72 CPU cores, and 768GB of RAM. Implementations are available under a CC-BY-SA license at TBD.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="5.1.">Results</head><p>Table <ref type="table" target="#tab_2">3</ref> illustrates our experimental results on both datasets. As evident from Table <ref type="table" target="#tab_2">3</ref>, traditional KGE methods perform poorly on both datasets, only slightly beating the random baseline on the macro F1 metric. Interestingly, both ComplEx and RotatE perform worse than TransE on both datasets. This finding is surprising as both ComplEx and RotatE are more expressive than TransE <ref type="bibr" target="#b20">[20]</ref>. However, when combined with MLP models, all exhibit significant performance boost, up to more than 100% in the case of RotatE. After this addition, we can see the same expressivity trend in the model results, i.e., the more powerful the model, the better the result. Moreover, the control "Random + MLP" experiment showcases very similar results to the random baseline, indicating the importance of both components for the hybrid model to perform well. Altogether, it is evident that the reasoning power of shallow traditional KGE models is not enough to capture the complexity of this task, and we require models with more reasoning power.</p><p>As for the MHLP method, in the transductive setting, it achieves 57.88 and 62.16 macro F1 scores on SciCiteorigin and SciCite resplit datasets, respectively. Moreover, its inductive results showcase the robustness of our approach in an extreme out-of-distribution setting, achieving 56.13 and 59.81 macro F1 scores. Compared to previously re-ported results <ref type="bibr" target="#b17">[17]</ref>, our model achieves superior performance to Glove-based models while slightly lagging behind Infersent-based models. Looking into the precision and recall comparison, our method has better precision scores on both transductive and inductive settings compared to all word embedding-based models; however, for recall, it performs better than Glove-based models and worse than the Infersent-based models which might stem from the imbalance in the links as illustrated by Figure <ref type="figure" target="#fig_3">3a</ref>. Further experimentation to address the class imbalance problem in future works might help improve the overall performance of MHLP. The significance of these results is that we show structural and relational information could be used to achieve relatively high performance without using textual information. Moreover, although our models underperform compared to language model-based approaches such as Structural Scaffolds and SciBERT, we showcase interesting future directions for combining graph-based and NLP-based methods.</p><p>Finally, the composite model denoted as SciBERT + MHLP in Table <ref type="table" target="#tab_2">3</ref>, achieves the best performance among all models, even beating the fine-tuned SciBERT. When considering MHLP's standalone performance, these results showcase the potential improvements that could be achieved through the use of structural information that is not available in citation phrases. The presented experiments are a stepping stone for better understanding and using the structural information at scale for citation intent classification.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.">Analysis</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.1.">Temporal Analysis</head><p>This analysis studies the relationship between the time that has passed since publication and citation intent. We hypothesize that a paper is more likely to be cited as "Result" or "Method" right after its publication, and as time passes, it will be more likely to be cited as "Background." If this is proven accurate, we could get a relatively strong signal from the temporal information for each citation. We plotted the years after publication against intent counts and ratios for all papers in the semantic scholar corpus to test our hypothesis. Figure <ref type="figure" target="#fig_3">3a</ref> and 3b illustrate the results of our analysis. As evident from these figures and contrary to our original hypothesis, we find out that the ratio of intent classes almost stays the same as time passes with insignificant fluctuations. As a result, using temporal information in our models is unlikely to provide any significant improvement. Note that these results are based on the weakly labeled links that we obtained from S2ORC. Consequently, these links are generated by another noisy model that could potentially be biased. Hence, it should not discourage further analysis or studies of temporal information for citation intent classification.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.2.">Mutual Information Analysis</head><p>In this analysis, we study the quality of the engineered features as described in Section 4.3 concerning the weakly labeled intent classes. To this end, we use the well-known mutual information (MI) <ref type="bibr" target="#b12">[12]</ref> measurement to quantify the importance of each feature. Formally, the MI between    </p><p>where 𝒴 is the value space for 𝑌 , 𝒳 is the value space for 𝑋, 𝑃𝑋,𝑌 is the joint probability distribution, and 𝑃𝑋 and 𝑃𝑌 are the marginal probability distributions. Note that MI is a non-negative value, and higher values indicate more correlation between the two random variables. For our analysis, we calculate MI for both sides of the 5,886 unique citation links in the SciCite resplit dataset. Moreover, to study these features in the graph context, we also calculate MI for the average of these features over the neighborhood of each publication, i.e., all citing and cited publications, from both sides of the citation links. Figures <ref type="figure" target="#fig_4">4a and 4b</ref> present the results of our experiments. As evident from these results, while the publication-averaged features generally show stronger connections to the target variable, the neighborhood-averaged features seem to show complementary connections, further emphasizing the importance of using both sets of features.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.3.">Feature Quality Analysis</head><p>In this analysis, we study the effect of normalization as described in Equations 6 and 7. To this end, we project the extracted features of the 5,886 unique citation links in the SciCite resplit dataset to a 2-dimensional space using t-SNE <ref type="bibr" target="#b24">[24]</ref>. Figure <ref type="figure" target="#fig_6">5a</ref> and 5b illustrate the projected space for the unnormalized and normalized features, respectively. As evident from Figure <ref type="figure" target="#fig_6">5a</ref>, it is challenging to distinguish different intent types in the unnormalized space. However, after normalization, as evident from Figure <ref type="figure" target="#fig_6">5b</ref>, we can see that the "Method" intention more or less creates a distinguishable cluster. This result shows that the use of normalization is potentially helpful for the model. Further studies on different types of normalization and their effects are left for future work.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="6.4.">Robustness Analysis</head><p>In this analysis, we focus on studying the robustness of our proposed graph-based method. To this end, we devise two ablation studies. In the first study, we randomly corrupt a percentage of the weak labels by replacing the correct label with a random label. This study aims to understand the model's resilience to noise better.</p><p>In the second study, we randomly remove a percentage of the weak labels. This study's idea is to understand better the effect of weak supervision on the model's performance. These studies are carried out by running the MHLP method in the transductive setting on both SciCiteorigin and SciCite resplit datasets. The feature vectors for the publications are calculated by counting the number of citations and intents. These vectors are normalized then using Equation 6 and 7. To analyze the relationship between the model's performance and the amount of available data, we create ten variations of the dataset by only using a portion of the available weak labels, varying from using all the available weak labels to only using 10% of them. Figure <ref type="figure" target="#fig_8">6a</ref> presents the result of this study.</p><p>As evident from Figure <ref type="figure" target="#fig_8">6a</ref>, the more weakly labeled links are available, the better our method performs. The other significant observation is the robustness of the model, even in the extreme scenario of having access to only 10% of the labels. Note that only 31.90% of links in the S2ROC have at least one weakly labeled intent, which means, even if the utilization percentage is 100%, only 31.90% citation links are weakly labeled.</p><p>Figure <ref type="figure" target="#fig_8">6b</ref> showcases the relationship between the model performance and the percentage of corrupted data. Following our intuition, the model's performance monotonically decreases as we add more noisy labels to the data. However, two interesting observations could be made from this figure. First, the performance of our method only drops less than five macro F1 scores when half (50%) of the weak labels are replaced with randomly assigned noisy labels. This observation shows that the proposed method is exceptionally resilient when faced with mistakes. Second, even when all the labels are replaced with random ones (100%), the model performs better than the random baselines. This observation indicates that the model is learning to make inferences based on purely structural information, which further solidifies our hypothesis regarding the importance of structural information.</p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head n="7.">Conclusions and Future Work</head><p>In this work, we first introduced an expansion to the Sci-Cite dataset by extracting scholarly information from the S2ORC dataset and creating an extended citation graph. Then, we gathered a large-scale weakly labeled dataset to augment the extracted citation graph with citation intents and create a multi-relational knowledge graph. Following this, we adapted the sentence-based intent classification into a citation-based link prediction task on graphs. We then introduced a set of engineered graph-based and citation-based features. Built on top of these features, we introduced a graph-based multi-hop reasoning approach for the newly introduced task. Our approach achieves 62.16 and 59.81 macro F1 scores in the transductive and inductive settings, respectively. The experimental results in the inductive setting further showcase the robustness of the proposed approach in the information-deprived outof-distribution environment. Compared to NLP-based models, we reached a comparable performance to, and in some cases outperform, the word embedding-based methods that rely on contextual sentences to make predictions. Moreover, with a composite model comprising our method as the graph encoder and the state-of-the-art NLP-based model as the text encoder, we outperformed all the other models we experimented with. These results further signify the strong signal in relational information and highlight the importance of future analysis and studies in this domain. Finally, our presented analyses further support our methodological choices.</p><p>For future works, one straightforward idea is to extend the knowledge graph with more scholarly information, such as authors, venues, and fields of study. There already exist some open repositories such as OpenAlex <ref type="bibr" target="#b15">[15]</ref> and Microsoft Academic Graph (MAG) <ref type="bibr" target="#b25">[25]</ref> that contain this information. Another direction is further investigation into the temporal signals. Last but not least, although we achieved an improved performance through a fusion of textual and structural information, more investigation and analysis could be done in this setting in future works.</p></div><figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_0"><head>Figure 1 :</head><label>1</label><figDesc>Figure 1: Overview of the extracted multi-hop KG. The set of 0-hop nodes 𝒱 0 includes all the orange nodes. The set of 1-hop nodes 𝒱 1 includes all the orange and blue nodes. Similarly, the graph could be expanded to include 𝑘-hop nodes 𝒱 𝑘 . The annotated set on each edge represents that specific link's intent. Specifically, the empty set denotes that the citation link has no intent label.</figDesc><graphic coords="2,312.79,84.19,183.02,133.53" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_1"><head>Figure 2 :</head><label>2</label><figDesc>Figure 2: Overview of the composite model. The model consists of two encoders for the citation phrase and the citation graph around the citation link. During the training phase, we freeze the SciBERT model in the first two epochs as a warm-up step for the graph encoder; then, we jointly train both encoders along with the final prediction module.</figDesc><graphic coords="6,89.29,84.19,416.68,128.79" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_2"><head></head><label></label><figDesc>(a) The number of different citation intents. (b) The percentage of different citation intents.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_3"><head>Figure 3 :</head><label>3</label><figDesc>Figure 3: The statistic of citation intent for all publications in the Semantic Scholar corpus. The temporal trends stay steady over time, suggesting a lack of information in the elapsed time from the time of publication to the time of citing.</figDesc><graphic coords="7,89.29,234.23,203.37,125.49" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_4"><head>Figure 4 :</head><label>4</label><figDesc>Figure 4: The calculated MI values for publication features and averaged neighborhood features. On average, the publication features show stronger connections to the target variable.</figDesc><graphic coords="8,302.62,277.88,203.36,176.90" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_5"><head></head><label></label><figDesc>(a) Features before normalization (b) Features after normalization</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_6"><head>Figure 5 :</head><label>5</label><figDesc>Figure 5: The t-SNE visualizations for the unnormalized and normalized features.</figDesc><graphic coords="9,89.29,276.04,203.36,124.33" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_7"><head></head><label></label><figDesc>(a) The percentage of utilized weak labels. (b) The percentage of corrupted data.</figDesc></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" xml:id="fig_8"><head>Figure 6 :</head><label>6</label><figDesc>Figure 6: The macro F1 score of MHLP (Transductive) on SciCite origin and SciCite resplit dataset</figDesc><graphic coords="9,89.29,417.17,203.37,127.35" type="bitmap" /></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_0"><head>Table 1</head><label>1</label><figDesc>The statistic of the SciCite dataset and reconstructed datasets.</figDesc><table><row><cell>Dataset</cell><cell>SciCite</cell><cell>SciCite origin</cell><cell>SciCite resplit</cell></row><row><cell>Level</cell><cell>Sentence</cell><cell>Link</cell><cell>Link</cell></row><row><cell># Samples</cell><cell>11,020</cell><cell>10,379</cell><cell>5,766</cell></row><row><cell># Train</cell><cell>8,243</cell><cell>7,602</cell><cell>4,122</cell></row><row><cell># Validation</cell><cell>916</cell><cell>916</cell><cell>822</cell></row><row><cell># Test</cell><cell>1,861</cell><cell>1,861</cell><cell>822</cell></row><row><cell>network [26]</cell><cell></cell><cell></cell><cell></cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_1"><head>Table 2</head><label>2</label><figDesc>Statistics of the extracted KGs along with the original S2ORC dataset.</figDesc><table><row><cell>Dataset</cell><cell cols="3"># Nodes # Citation Links # Background</cell><cell># Method</cell><cell cols="2"># Result Weak Labels</cell></row><row><cell>Zero-Hop (𝒢 0 )</cell><cell>13,011</cell><cell>10,733</cell><cell>5,479</cell><cell>4,403</cell><cell>1,335</cell><cell>79.04%</cell></row><row><cell>One-Hop (𝒢 1 )</cell><cell>5,862,261</cell><cell>119,776,090</cell><cell>39,202,086</cell><cell cols="2">16,830,665 16,830,665</cell><cell>43.18%</cell></row><row><cell>Two-Hop (𝒢 2 )</cell><cell>57,535,880</cell><cell>1,621,293,902</cell><cell cols="3">467,860,523 121,877,053 35,283,718</cell><cell>34.41%</cell></row><row><cell>S2ORC</cell><cell>206,159,629</cell><cell>2,495,513,737</cell><cell cols="3">643,955,457 169,472,164 45,779,793</cell><cell>31.90%</cell></row></table></figure>
<figure xmlns="http://www.tei-c.org/ns/1.0" type="table" xml:id="tab_2"><head>Table 3</head><label>3</label><figDesc>Intent classification results on SciCite origin and SciCite resplit datasets. All the metrics are macro averaged. Bold values represent the highest performance within the metric and dataset scope.</figDesc><table><row><cell>SciCite origin</cell><cell>SciCite resplit</cell></row></table></figure>
		</body>
		<back>

			<div type="acknowledgement">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>Acknowledgments</head><p>This work was funded by the Defense Advanced Research Projects Agency with award W911NF-19-20271 and with support from a Keston Exploratory Research Award.</p></div>
			</div>

			<div type="annex">
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A. Hyperparameters</head></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A.1. Knowledge Graph Embedding</head><p>We use a randomized search to tune our models and find near-optimal hyperparameters using the following ranges: embedding dimensions ∈ {50, 100, 200}, learning rate ∈ {0.03, 0.1, 0.3}, regularization coefficient ∈ {0.0, 1e-9, 1e-8, 1e- </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A.2. Multilayer Perceptron</head><p>To simplify the model tuning process, we find the optimal hyperparameters of "ComplEx + MLP" on SciCiteorigin using grid search and reuse them for the rest of our experiments. Specifically, we run a grid search over the following ranges: number of layers ∈ {0, 1, 2, 3}, dropout ∈ {0, 0. </p></div>
<div xmlns="http://www.tei-c.org/ns/1.0"><head>A.3. Multi-Hop Link Prediction</head><p>We run a grid search over the following ranges: number of layers ∈ {0, 1, 2, 3}, dimension ∈ {10, 50, 100, 200}, learning rate ∈ {0.03, 0.01, 0.003, 0.001}, The optimal hyperparameters are as follows: number of layers = 1, dimension = 100, learning rate = 0.01. We use Adam as the optimizer through the tuning process.</p><p>We use a randomized search to tune our models and find near-optimal hyperparameters using the following ranges: embedding dimensions ∈ {50, 100, 200}, learning rate ∈ {0.03, 0.1, 0.3}, regularization coefficient ∈ {0.0, 1e-9, 1e-8, 1e-7, 1e-6, 1e-5}, number of negative samples ∈ {64, 128, 256, 512, 1024}, 𝛼 ∈ {0.25, 0.5, 1}, 𝛾 ∈ {6, 12, 24}. Note that 𝛼 and 𝛾 are the adversarial temperature and the margin value (RotatEonly), respectively.</p></div>			</div>
			<div type="references">

				<listBibl>

<biblStruct xml:id="b0">
	<analytic>
		<title level="a" type="main">SciB-ERT: A Pretrained Language Model for Scientific Text</title>
		<author>
			<persName><forename type="first">Iz</forename><surname>Beltagy</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Kyle</forename><surname>Lo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Arman</forename><surname>Cohan</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/D19-1371</idno>
		<ptr target="https://doi.org/10.18653/v1/D19-1371" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). Association for Computational Linguistics</title>
				<meeting>the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP). Association for Computational Linguistics<address><addrLine>Hong Kong, China</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2019">2019</date>
			<biblScope unit="page" from="3615" to="3620" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b1">
	<monogr>
		<title/>
		<author>
			<persName><forename type="first">Antoine</forename><surname>Bordes</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Nicolas</forename><surname>Usunier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alberto</forename><surname>Garcia-Durán</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jason</forename><surname>Weston</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Oksana</forename><surname>Yakhnenko</surname></persName>
		</author>
		<imprint/>
	</monogr>
</biblStruct>

<biblStruct xml:id="b2">
	<analytic>
		<title level="a" type="main">Translating Embeddings for Modeling Multi-Relational Data</title>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 26th International Conference on Neural Information Processing Systems -Volume 2</title>
				<meeting>the 26th International Conference on Neural Information Processing Systems -Volume 2<address><addrLine>Lake Tahoe, Nevada; Red Hook, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Curran Associates Inc</publisher>
			<biblScope unit="page" from="2787" to="2795" />
		</imprint>
	</monogr>
	<note>NIPS&apos;13</note>
</biblStruct>

<biblStruct xml:id="b3">
	<analytic>
		<title level="a" type="main">What do citation counts measure? A review of studies on citing behavior</title>
		<author>
			<persName><forename type="first">Lutz</forename><surname>Bornmann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hans-Dieter</forename><surname>Daniel</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">J. Documentation</title>
		<imprint>
			<biblScope unit="volume">64</biblScope>
			<biblScope unit="page" from="45" to="80" />
			<date type="published" when="2008">2008. 2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b4">
	<analytic>
		<title level="a" type="main">Structural Scaffolds for Citation Intent Classification in Scientific Publications</title>
		<author>
			<persName><forename type="first">Arman</forename><surname>Cohan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Waleed</forename><surname>Ammar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Madeleine</forename><surname>Van Zuylen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Field</forename><surname>Cady</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/N19-1361</idno>
		<ptr target="https://doi.org/10.18653/v1/N19-1361" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</title>
		<title level="s">Long and Short Papers</title>
		<meeting>the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies<address><addrLine>Minneapolis, Minnesota</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computational Linguistics</publisher>
			<date type="published" when="2019">2019</date>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="page" from="3586" to="3596" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b5">
	<analytic>
		<title level="a" type="main">Scientific Article Summarization Using Citation-Context and Article&apos;s Discourse Structure</title>
		<author>
			<persName><forename type="first">Arman</forename><surname>Cohan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Nazli</forename><surname>Goharian</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/D15-1045</idno>
		<ptr target="https://doi.org/10.18653/v1/D15-1045" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing</title>
				<meeting>the 2015 Conference on Empirical Methods in Natural Language Processing<address><addrLine>Lisbon, Portugal</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="390" to="400" />
		</imprint>
	</monogr>
	<note>Association for Computational Linguistics</note>
</biblStruct>

<biblStruct xml:id="b6">
	<analytic>
		<title level="a" type="main">Structured Citation Trend Prediction Using Graph Neural Networks</title>
		<author>
			<persName><forename type="first">Daniel</forename><surname>Cummings</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Marcel</forename><surname>Nassar</surname></persName>
		</author>
		<ptr target="http://dblp.uni-trier.de/db/conf/icassp/icassp2020.html#CummingsN20" />
	</analytic>
	<monogr>
		<title level="m">ICASSP. IEEE</title>
				<meeting><address><addrLine>Barcelona, Spain</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="3897" to="3901" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b7">
	<analytic>
		<title level="a" type="main">Automated Classification of Citations Using Linguistic Semantic Grammars</title>
		<author>
			<persName><forename type="first">M</forename><forename type="middle">A</forename><surname>Garzone</surname></persName>
		</author>
		<ptr target="https://books.google.com/books?id=V-bwSgAACAAJ" />
	</analytic>
	<monogr>
		<title level="m">Thesis</title>
				<meeting><address><addrLine>London, Canada</addrLine></address></meeting>
		<imprint>
			<date type="published" when="1997">1997</date>
		</imprint>
		<respStmt>
			<orgName>M.Sc.)-University of Western Ontario</orgName>
		</respStmt>
	</monogr>
</biblStruct>

<biblStruct xml:id="b8">
	<analytic>
		<title level="a" type="main">Inductive Representation Learning on Large Graphs</title>
		<author>
			<persName><forename type="first">William</forename><forename type="middle">L</forename><surname>Hamilton</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rex</forename><surname>Ying</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jure</forename><surname>Leskovec</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 31st International Conference on Neural Information Processing Systems</title>
				<meeting>the 31st International Conference on Neural Information Processing Systems<address><addrLine>Long Beach, California, USA; Red Hook, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Curran Associates Inc</publisher>
			<date type="published" when="2017">2017</date>
			<biblScope unit="page" from="1025" to="1035" />
		</imprint>
	</monogr>
	<note>NIPS&apos;17</note>
</biblStruct>

<biblStruct xml:id="b9">
	<analytic>
		<title level="a" type="main">Survey about citation context analysis: Tasks, techniques, and resources</title>
		<author>
			<persName><forename type="first">Myriam</forename><surname>Hernández</surname></persName>
		</author>
		<author>
			<persName><forename type="first">-</forename><surname>Alvarez</surname></persName>
		</author>
		<author>
			<persName><forename type="first">José</forename><forename type="middle">M</forename><surname>Gomez</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Natural Language Engineering</title>
		<imprint>
			<biblScope unit="volume">22</biblScope>
			<biblScope unit="issue">3</biblScope>
			<biblScope unit="page" from="327" to="349" />
			<date type="published" when="2016">2016. 2016</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b10">
	<analytic>
		<title level="a" type="main">Measuring the Evolution of a Scientific Field through Citation Frames</title>
		<author>
			<persName><forename type="first">David</forename><surname>Jurgens</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Srijan</forename><surname>Kumar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Raine</forename><surname>Hoover</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Dan</forename><surname>Mcfarland</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Dan</forename><surname>Jurafsky</surname></persName>
		</author>
		<idno type="DOI">10.1162/tacl_a_00028</idno>
		<ptr target="https://doi.org/10.1162/tacl_a_00028" />
	</analytic>
	<monogr>
		<title level="j">Transactions of the Association for Computational Linguistics</title>
		<imprint>
			<biblScope unit="volume">6</biblScope>
			<biblScope unit="page" from="391" to="406" />
			<date type="published" when="2018">2018. 2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b11">
	<analytic>
		<title level="a" type="main">Semi-Supervised Classification with Graph Convolutional Networks</title>
		<author>
			<persName><forename type="first">Thomas</forename><forename type="middle">N</forename><surname>Kipf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Max</forename><surname>Welling</surname></persName>
		</author>
		<ptr target="https://openreview.net/forum?id=SJU4ayYgl" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 5th International Conference on Learning Representations (ICLR &apos;17). OpenReview.net</title>
				<meeting>the 5th International Conference on Learning Representations (ICLR &apos;17). OpenReview.net<address><addrLine>Palais des Congrès Neptune, Toulon, France</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2014">2017. 14</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b12">
	<analytic>
		<title level="a" type="main">Estimating mutual information</title>
		<author>
			<persName><forename type="first">Alexander</forename><surname>Kraskov</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Harald</forename><surname>Stögbauer</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Peter</forename><surname>Grassberger</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Physical review E</title>
		<imprint>
			<biblScope unit="volume">69</biblScope>
			<biblScope unit="issue">6</biblScope>
			<biblScope unit="page">66138</biblScope>
			<date type="published" when="2004">2004. 2004</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b13">
	<analytic>
		<title level="a" type="main">S2ORC: The Semantic Scholar Open Research Corpus</title>
		<author>
			<persName><forename type="first">Kyle</forename><surname>Lo</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lucy</forename><forename type="middle">Lu</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mark</forename><surname>Neumann</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rodney</forename><surname>Kinney</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Daniel</forename><surname>Weld</surname></persName>
		</author>
		<idno type="DOI">10.18653/v1/2020.acl-main.447</idno>
		<ptr target="https://doi.org/10.18653/v1/2020.acl-main.447" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics</title>
				<meeting>the 58th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics</meeting>
		<imprint>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="4969" to="4983" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b14">
	<analytic>
		<title level="a" type="main">Automatic Differentiation in PyTorch</title>
		<author>
			<persName><forename type="first">Adam</forename><surname>Paszke</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sam</forename><surname>Gross</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Soumith</forename><surname>Chintala</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Gregory</forename><surname>Chanan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Edward</forename><surname>Yang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zachary</forename><surname>Devito</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zeming</forename><surname>Lin</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Alban</forename><surname>Desmaison</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Luca</forename><surname>Antiga</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Adam</forename><surname>Lerer</surname></persName>
		</author>
		<ptr target="https://openreview.net/forum?id=BJJsrmfCZ" />
	</analytic>
	<monogr>
		<title level="m">NIPS 2017 Workshop on Autodiff. OpenReview.net</title>
				<meeting><address><addrLine>Long Beach, California, USA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2017">2017</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b15">
	<monogr>
		<author>
			<persName><forename type="first">Jason</forename><surname>Priem</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Heather</forename><surname>Piwowar</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Richard</forename><surname>Orr</surname></persName>
		</author>
		<idno type="arXiv">arXiv:2205.01833abs/2205.01833</idno>
		<title level="m">OpenAlex: A fully-open index of scholarly works, authors, venues, institutions, and concepts</title>
				<imprint>
			<date type="published" when="2022">2022. 2022</date>
		</imprint>
	</monogr>
	<note type="report_type">arXiv preprint</note>
	<note>5 pages</note>
</biblStruct>

<biblStruct xml:id="b16">
	<monogr>
		<title level="m" type="main">Citation context analysis for information retrieval</title>
		<author>
			<persName><forename type="first">Anna</forename><surname>Ritchie</surname></persName>
		</author>
		<imprint>
			<date type="published" when="2009">2009</date>
		</imprint>
		<respStmt>
			<orgName>University of Cambridge, Computer Laboratory</orgName>
		</respStmt>
	</monogr>
	<note type="report_type">Technical Report</note>
</biblStruct>

<biblStruct xml:id="b17">
	<analytic>
		<title level="a" type="main">Citation intent classification using word embedding</title>
		<author>
			<persName><forename type="first">Muhammad</forename><surname>Roman</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Abdul</forename><surname>Shahid</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Shafiullah</forename><surname>Khan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Anis</forename><surname>Koubaa</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Lisu</forename><surname>Yu</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Ieee Access</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="9982" to="9995" />
			<date type="published" when="2021">2021. 2021</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b18">
	<analytic>
		<title level="a" type="main">Modeling Relational Data with Graph Convolutional Networks</title>
		<author>
			<persName><forename type="first">Michael</forename><surname>Schlichtkrull</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Thomas</forename><forename type="middle">N</forename><surname>Kipf</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Peter</forename><surname>Bloem</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Rianne</forename><surname>Van Den</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Ivan</forename><surname>Berg</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Max</forename><surname>Titov</surname></persName>
		</author>
		<author>
			<persName><surname>Welling</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">The Semantic Web</title>
				<editor>
			<persName><forename type="first">Aldo</forename><surname>Gangemi</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Roberto</forename><surname>Navigli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Maria-Esther</forename><surname>Vidal</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Pascal</forename><surname>Hitzler</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Raphaël</forename><surname>Troncy</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Laura</forename><surname>Hollink</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Anna</forename><surname>Tordai</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Mehwish</forename><surname>Alam</surname></persName>
		</editor>
		<meeting><address><addrLine>Cham</addrLine></address></meeting>
		<imprint>
			<publisher>Springer International Publishing</publisher>
			<date type="published" when="2018">2018</date>
			<biblScope unit="page" from="593" to="607" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b19">
	<analytic>
		<title level="a" type="main">Characterizing highly cited method and non-method papers using citation contexts: The role of uncertainty</title>
		<author>
			<persName><forename type="first">Henry</forename><surname>Small</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Journal of Informetrics</title>
		<imprint>
			<biblScope unit="volume">12</biblScope>
			<biblScope unit="issue">2</biblScope>
			<biblScope unit="page" from="461" to="480" />
			<date type="published" when="2018">2018. 2018</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b20">
	<analytic>
		<title level="a" type="main">RotatE: Knowledge Graph Embedding by Relational Rotation in Complex Space</title>
		<author>
			<persName><forename type="first">Zhiqing</forename><surname>Sun</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zhi-Hong</forename><surname>Deng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jian-Yun</forename><surname>Nie</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jian</forename><surname>Tang</surname></persName>
		</author>
		<ptr target="http://dblp.uni-trier.de/db/conf/iclr/iclr2019.html#SunDNT19" />
	</analytic>
	<monogr>
		<title level="m">ICLR (Poster). OpenReview.net</title>
				<meeting><address><addrLine>New Orleans, LA</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2018">2019. 18</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b21">
	<analytic>
		<title level="a" type="main">An annotation scheme for citation function</title>
		<author>
			<persName><forename type="first">Simone</forename><surname>Teufel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Advaith</forename><surname>Siddharthan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Dan</forename><surname>Tidhar</surname></persName>
		</author>
		<ptr target="https://aclanthology.org/W06-1312" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 7th SIGdial Workshop on Discourse and Dialogue. Association for Computational Linguistics</title>
				<meeting>the 7th SIGdial Workshop on Discourse and Dialogue. Association for Computational Linguistics<address><addrLine>Sydney, Australia</addrLine></address></meeting>
		<imprint>
			<date type="published" when="2006">2006</date>
			<biblScope unit="page" from="80" to="87" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b22">
	<analytic>
		<title level="a" type="main">Complex Embeddings for Simple Link Prediction</title>
		<author>
			<persName><forename type="first">Théo</forename><surname>Trouillon</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Johannes</forename><surname>Welbl</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Sebastian</forename><surname>Riedel</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Éric</forename><surname>Gaussier</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Guillaume</forename><surname>Bouchard</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 33rd International Conference on International Conference on Machine Learning -Volume 48 (ICML&apos;16</title>
				<meeting>the 33rd International Conference on International Conference on Machine Learning -Volume 48 (ICML&apos;16<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>JMLR.org</publisher>
			<date type="published" when="2016">2016</date>
			<biblScope unit="page" from="2071" to="2080" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b23">
	<analytic>
		<title level="a" type="main">Identifying Meaningful Citations</title>
		<author>
			<persName><forename type="first">Marco</forename><surname>Valenzuela</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Vu</forename><surname>Ha</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Oren</forename><surname>Etzioni</surname></persName>
		</author>
		<ptr target="http://www.aaai.org/Library/Workshops/ws15-13.php" />
	</analytic>
	<monogr>
		<title level="m">Scholarly Big Data: AI Perspectives, Challenges, and Ideas, Papers from the 2015 AAAI Workshop (Technical Report</title>
				<editor>
			<persName><forename type="first">Cornelia</forename><surname>Caragea</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">C</forename><forename type="middle">Lee</forename><surname>Giles</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Narayan</forename><surname>Bhamidipati</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Doina</forename><surname>Caragea</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Sujatha</forename><surname>Das Gollapalli</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Saurabh</forename><surname>Kataria</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Huan</forename><surname>Liu</surname></persName>
		</editor>
		<editor>
			<persName><forename type="first">Feng</forename><surname>Xia</surname></persName>
		</editor>
		<meeting><address><addrLine>Menlo Park, CA</addrLine></address></meeting>
		<imprint>
			<publisher>AAAI Press</publisher>
			<date type="published" when="2015">2015</date>
			<biblScope unit="page" from="21" to="26" />
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b24">
	<analytic>
		<title level="a" type="main">Visualizing Data using t-SNE</title>
		<author>
			<persName><forename type="first">Laurens</forename><surname>Van Der Maaten</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Geoffrey</forename><surname>Hinton</surname></persName>
		</author>
		<ptr target="http://jmlr.org/papers/v9/vandermaaten08a.html" />
	</analytic>
	<monogr>
		<title level="j">Journal of Machine Learning Research</title>
		<imprint>
			<biblScope unit="volume">9</biblScope>
			<biblScope unit="page" from="2579" to="2605" />
			<date type="published" when="2008">2008. 2008</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b25">
	<analytic>
		<title level="a" type="main">Microsoft academic graph: When experts are not enough</title>
		<author>
			<persName><forename type="first">Kuansan</forename><surname>Wang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zhihong</forename><surname>Shen</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Chiyuan</forename><surname>Huang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Chieh-Han</forename><surname>Wu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Yuxiao</forename><surname>Dong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Anshul</forename><surname>Kanakia</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="j">Quantitative Science Studies</title>
		<imprint>
			<biblScope unit="volume">1</biblScope>
			<biblScope unit="issue">1</biblScope>
			<biblScope unit="page" from="396" to="413" />
			<date type="published" when="2020">2020. 2020</date>
		</imprint>
	</monogr>
</biblStruct>

<biblStruct xml:id="b26">
	<analytic>
		<title level="a" type="main">Identifying Referential Intention with Heterogeneous Contexts</title>
		<author>
			<persName><forename type="first">Wenhao</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Mengxia</forename><surname>Yu</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Tong</forename><surname>Zhao</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Meng</forename><surname>Jiang</surname></persName>
		</author>
		<idno type="DOI">10.1145/3366423.3380175</idno>
		<ptr target="https://doi.org/10.1145/3366423.3380175" />
	</analytic>
	<monogr>
		<title level="m">Proceedings of The Web Conference 2020</title>
				<meeting>The Web Conference 2020<address><addrLine>Taipei, Taiwan; New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="962" to="972" />
		</imprint>
	</monogr>
	<note>WWW &apos;20</note>
</biblStruct>

<biblStruct xml:id="b27">
	<analytic>
		<title level="a" type="main">DGL-KE: Training Knowledge Graph Embeddings at Scale</title>
		<author>
			<persName><forename type="first">Da</forename><surname>Zheng</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Xiang</forename><surname>Song</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Chao</forename><surname>Ma</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zeyuan</forename><surname>Tan</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zihao</forename><surname>Ye</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Jin</forename><surname>Dong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Hao</forename><surname>Xiong</surname></persName>
		</author>
		<author>
			<persName><forename type="first">Zheng</forename><surname>Zhang</surname></persName>
		</author>
		<author>
			<persName><forename type="first">George</forename><surname>Karypis</surname></persName>
		</author>
	</analytic>
	<monogr>
		<title level="m">Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR &apos;20)</title>
				<meeting>the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR &apos;20)<address><addrLine>New York, NY, USA</addrLine></address></meeting>
		<imprint>
			<publisher>Association for Computing Machinery</publisher>
			<date type="published" when="2020">2020</date>
			<biblScope unit="page" from="739" to="748" />
		</imprint>
	</monogr>
</biblStruct>

				</listBibl>
			</div>
		</back>
	</text>
</TEI>
