<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Evaluating Pretrained Transformer Models for Citation Recommendation</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Rodrigo Nogueira</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
          <xref ref-type="aff" rid="aff5">5</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Zhiying Jiang</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Kyunghyun Cho</string-name>
          <xref ref-type="aff" rid="aff0">0</xref>
          <xref ref-type="aff" rid="aff1">1</xref>
          <xref ref-type="aff" rid="aff2">2</xref>
          <xref ref-type="aff" rid="aff4">4</xref>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Jimmy Lin</string-name>
          <xref ref-type="aff" rid="aff3">3</xref>
        </contrib>
        <aff id="aff0">
          <label>0</label>
          <institution>CIFAR Azrieli Global Scholar</institution>
        </aff>
        <aff id="aff1">
          <label>1</label>
          <institution>Center for Data Science, New York University</institution>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff2">
          <label>2</label>
          <institution>Courant Institute of Mathematical Sciences, New York University</institution>
          ,
          <country country="US">USA</country>
        </aff>
        <aff id="aff3">
          <label>3</label>
          <institution>David R. Cheriton School of Computer Science, University of Waterloo</institution>
        </aff>
        <aff id="aff4">
          <label>4</label>
          <institution>Facebook AI Research</institution>
        </aff>
        <aff id="aff5">
          <label>5</label>
          <institution>Tandon School of Engineering, New York University</institution>
          ,
          <country country="US">USA</country>
        </aff>
      </contrib-group>
      <pub-date>
        <year>2020</year>
      </pub-date>
      <fpage>89</fpage>
      <lpage>100</lpage>
      <abstract>
        <p>Citation recommendation systems for the scienti c literature, to help authors nd papers that should be cited, have the potential to speed up discoveries and uncover new routes for scienti c exploration. We treat this task as a ranking problem, which we tackle with a twostage approach: candidate generation followed by re-ranking. Within this framework, we adapt to the scienti c domain a proven combination based on \bag of words" retrieval followed by re-scoring with a BERT model. We experimentally show the e ects of domain adaptation, both in terms of pretraining on in-domain data and exploiting in-domain vocabulary. In addition, we evaluate eleven pretrained transformer models and analyze some unexpected failure cases. On three di erent collections from di erent scienti c disciplines, our models perform close to or at the state of the art in the citation recommendation task.</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>Introduction</title>
      <p>
        The volume of scienti c publications is growing at an incredible rate. For
example, over 900,000 papers are added per year to MEDLINE, a database of the life
sciences and biomedical literature.1 A recent study estimates that 3M papers
are published annually in the English language, with a growth rate of 3{5% per
year [
        <xref ref-type="bibr" rid="ref18">18</xref>
        ]. This ood of information has made it nearly impossible for researchers
to keep abreast of discoveries and innovations, both in their speci c sub- eld as
well as more broadly. Furthermore, there is an overwhelming amount of
material that a scientist entering a new eld of study needs to read before becoming
familiarized with common concepts, methods, and other foundations.
      </p>
      <p>
        A number of tools have come along to help researchers cope with this
deluge. For example, keyword-based literature search engines (Google Scholar,
Microsoft Academic, PubMed, and Semantic Scholar) and citation recommendation
tools [
        <xref ref-type="bibr" rid="ref14 ref2 ref21 ref27 ref5">5, 2, 27, 21, 14</xref>
        ] help scientists nd relevant articles, often exploiting
citation networks to identify what's important in a particular eld. Methods to
automatically populate scienti c knowledge bases [
        <xref ref-type="bibr" rid="ref12 ref34 ref35">12, 34, 35</xref>
        ] form another broad
approach to tackling this challenge.
      </p>
      <p>
        In this work, we investigate the potential of deep pretrained transformer
models such as BERT [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] and large scienti c datasets such as Open Research [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ]
to improve scienti c search tools. More concretely, we tackle the task of
scienti c literature recommendation, where a paper (title and abstract) is given as
a query, and the system's task is to nd papers that should be cited. We use a
standard keyword search engine (based on inverted indexes) with BM25
ranking [
        <xref ref-type="bibr" rid="ref33">33</xref>
        ] to initially retrieve candidate documents and evaluate various pretrained
transformer models as re-rankers.
      </p>
      <p>
        We nd that this simple pipeline is more e ective than previous cluster-based
methods [
        <xref ref-type="bibr" rid="ref32 ref4">32, 4</xref>
        ]. To summarize, our main contributions are as follows:
{ We evaluate eleven pretrained ranking models and nd that pretraining on
the target domain and using domain-speci c vocabulary leads to large
improvements over a general-purpose model.
{ We nd that despite the e ectiveness of the pretrained transformer models
as query{document relevance estimators, they perform poorly when the term
overlap between the query and candidate documents is low. To address this
issue, we train with more query{candidate pairs that have low term overlap,
but interestingly, such a model performs poorly, even on the training set (see
Section 5.2).
{ Contrary to our expectation given the symmetric nature of query and
candidate documents, we nd that query terms are more important than candidate
document terms for relevance estimation (see Section 5.3).
2
      </p>
    </sec>
    <sec id="sec-2">
      <title>Related Work</title>
      <p>
        Most early methods for scienti c literature search and recommendation take
advantage of keyword-based retrieval [
        <xref ref-type="bibr" rid="ref13 ref22">13, 22</xref>
        ]. These techniques su er from the
term mismatch problem, which is common in \bag-of-words" retrieval methods,
but the issue is aggravated by the diversity of scienti c vocabulary [
        <xref ref-type="bibr" rid="ref17 ref29 ref8">17, 8, 29</xref>
        ].
As the number of users grows, popular search engines can exploit interaction
signals to learn better ranking models [
        <xref ref-type="bibr" rid="ref10 ref11 ref28">28, 11, 10</xref>
        ]. However, the reported gains
are relatively small compared to classic ranking methods such as BM25.
      </p>
      <p>
        Another common approach in scienti c recommendation systems is
collaborative ltering [
        <xref ref-type="bibr" rid="ref24 ref27 ref6">27, 24, 6</xref>
        ]. These methods typically su er from the cold-start
problem, in which there is not enough evidence about new items (or users) to
make predictions accurately.
      </p>
      <p>
        More recently, cluster-based methods have started to become competitive
with traditional retrieval-based methods in this task. Kanakia et al. [
        <xref ref-type="bibr" rid="ref19">19</xref>
        ] cluster
papers based on their word embedding representation and use co-citations to
alleviate the cold-start problem. However, they perform human evaluations on
a private dataset, which excludes an empirical comparison to our approach.
      </p>
      <p>
        Perhaps closest to our work is Eto [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], who uses a combination of
proximity measures from the graph of co-citations to score candidate documents. The
edges in the graph are weighted by the distance in which two citations occur in
the citing document. This method requires access to the full text of the citing
document, which is often not available (for example, due to paywalled content).
Our method, on the other hand, predicts citations using only article abstracts,
which are widely available in scienti c corpora.
      </p>
      <p>
        The methods described so far and our work fall in the category of global
methods, which aim at recommending citations for the entire paper. Another
category comprises local methods, which aim at recommending citations for a
speci c sentence or paragraph in the document [
        <xref ref-type="bibr" rid="ref14 ref15 ref16 ref26">14, 26, 15, 16</xref>
        ]. We do not
compare our method to these as we do not assume access to the full text.
3
      </p>
    </sec>
    <sec id="sec-3">
      <title>Methods</title>
      <p>This work tackles the task of citation recommendation: given a partially written
paper, the system's task is to return all papers that should be cited in it. The
input query q is the title and abstract of a paper (and not the full text). We argue
that this assumption is crucial to building a useful tool as authors might desire
recommendations of relevant citations prior to writing most of their paper.</p>
      <p>
        Our method comprises two phases, Retrieval and Ranking. In the rst phase,
the top-k papers D are retrieved by a keyword search engine when queried with
query q. In the second phase, we compute the probability p(djq) of each paper
d 2 D being relevant to q. For this, we use a BERT [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ] re-ranker model based
on Nogueira and Cho [
        <xref ref-type="bibr" rid="ref30">30</xref>
        ]. Using the same notation as Devlin et al., we feed the
query tokens as sequence A and the candidate paper tokens as sequence B.
      </p>
      <p>In our setup, both the query and the candidate are the concatenation of the
title and abstract of each paper, resulting in an input sequence that is often
longer than the maximum tokens allowed by the model (typically 512 tokens).
To handle this, we devote 256 tokens for the query and 256 for the candidate,
truncating as necessary. At inference time, we use the model as a binary classi er:
we feed the [CLS] token to a single layer neural network to obtain p(djq). The
output of our method is a list of papers D ranked by p(djq). Training details are
provided in Section 4.2.
4
4.1</p>
    </sec>
    <sec id="sec-4">
      <title>Experimental Setup</title>
      <sec id="sec-4-1">
        <title>Datasets</title>
        <p>
          Open Research. We train and evaluate our models on the Open Research
corpus [
          <xref ref-type="bibr" rid="ref1">1</xref>
          ],2 comprising 7.2M computer science and biomedical paper abstracts and
their references. We closely follow the data processing steps from Bhagavatula
2 https://s3-us-west-2.amazonaws.com/ai2-s2-research-public/open-corpus/
2017-02-21/papers-2017-02-21.zip
et al. [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] to create the training, development, and test sets. In more detail, we
sort papers by publication year and use the oldest 80% for training (1991{2014),
the next 10% for development (2014{2015), and the most recent 10% for testing
(2015{2016). Since the development and test sets are too large (400k+ papers),
we randomly sample 20k examples from each set. We remove papers that do
not cite any other paper or that have no year of publication. Finally, we remove
citations of papers that are not in the corpus or whose year of publication is later
than that of the citing paper. Table 1 shows the statistics of the nal dataset
after all processing steps.
        </p>
        <p>
          Note that although our dataset statistics do not match those reported in
Bhagavatula et al. [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], they match the output of the evaluation script provided
by the authors.3 The di erence is that the authors report statistics before the
ltering steps (e.g., removing papers without references). Thus, our corpus and
dataset splits match exactly and thus our results are comparable.
DBLP and PubMed. The DBLP and PubMed datasets were introduced by
Ren et al. [
          <xref ref-type="bibr" rid="ref32">32</xref>
          ] and comprise papers from computer science and biomedicine,
respectively. We apply the same data processing steps from Bhagavatula et al.,
and the resulting dataset statistics are summarized in Table 1.
        </p>
        <p>Once processed in the manner described above, the citations within each paper
serve as the ground truth for that paper. That is, using a speci c paper as a
query, the perfect results set comprises the actual citations in that paper.</p>
        <p>When evaluating our method on DBLP and PubMed, we use models trained
on Open Research's training set as this yields better results than training on the
much smaller DBLP and PubMed training sets. To avoid leaking training data
into the evaluation sets, we use the following method to remove documents in
3 https://github.com/allenai/citeomatic/blob/master/citeomatic/scripts/
evaluate.py
Open Research's training set that appear in the development and test sets of
PubMed and DBLP: We remove special characters from the title and use Jaccard
similarity (on unigrams) to calculate the closeness of two documents, ltering
with a threshold of 0.7. This method results in approximately half of the papers
in the development and test sets of PubMed and DBLP being removed from the
training set of Open Research.
4.2</p>
      </sec>
      <sec id="sec-4-2">
        <title>Re-ranker Training</title>
        <p>
          To obtain the positive and negative examples used to train our binary classi
cation models, we retrieve the top 10 papers for each query (title + abstract)
using the Anserini IR toolkit4 [
          <xref ref-type="bibr" rid="ref36 ref37">36, 37</xref>
          ] with BM25 ranking. Among these,
approximately 6% on average are relevant papers (positive examples). We do not
balance positive and negative examples; see additional discussions about this
decision in Section 5.2.
        </p>
        <p>Starting with a pretrained BERT model, we ne-tune it to our task using
cross-entropy loss:</p>
        <p>L =</p>
        <p>X log(p(dj jq))
j2Jpos</p>
        <p>X log(1
j2Jneg
p(dj jq));
(1)
where Jpos and Jneg are the indexes of the relevant and non-relevant papers
and p(dj jq) is the relevance probability the model assigns to the j-th paper. We
examine several BERT variants, detailed in Section 5.1.</p>
        <p>
          All models are ne-tuned using Google's TPUs v3-8 with a batch size of
128 (128 sequences 512 tokens = 65,536 tokens/batch) for 300k iterations,
which takes approximately three days. This corresponds to training on 38.4M
(300k 128) query{candidate pairs, or 1.1 epochs. We do not see any
improvements in the development set when training for another 700k iterations, which
is equivalent to 3.8 epochs. We use Adam [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ] with the initial learning rate set
to 3 10 6, 1 = 0:9, 2 = 0:999, L2 weight decay of 0.01, learning rate
warmup over the rst 10,000 steps, and linear decay of the learning rate. We use a
dropout probability of 0:1 in all layers.
4.3
        </p>
      </sec>
      <sec id="sec-4-3">
        <title>Inference and Metrics</title>
        <p>
          At inference time, we rst retrieve the top 1000 candidate documents with the
title and abstract as the query using BM25 ranking in Anserini. These
documents are further re-ranked with one of the variants of the ne-tuned BERT
models (see Section 5.1 for more details). Following Bhagavatula et al. [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ], we
evaluate the results using F1 of the top 20 retrieved papers (F1@20) and Mean
Reciprocal Ranking (MRR) of the top 1000 retrieved papers. We additionally
report Recall@1000 (R@1000) to assess the e ectiveness of our keyword search
in isolation, which provides an upper bound on re-ranking e ectiveness.
        </p>
        <sec id="sec-4-3-1">
          <title>4 http://anserini.io/</title>
          <p>Our main results are shown in Table 2 with SciBERT-Large as the ranking
model, selected based on the experiments in Section 5.1. On the Open Research
dataset, our best con guration (BM25 + SciBERT-Large) improves upon the
best previous result in terms of both F1@20 and MRR. On the smaller DBLP
and PubMed datasets, our method is on par with the state of the art. Note that
our BERT-based models are trained only on Open Research as we achieve better
results than training on the smaller datasets.</p>
          <p>Interestingly, our baseline BM25 implementation using Anserini out of the
box, denoted \BM25 (Anserini)" in Table 2, is 3{7 points higher in F1@20 than
the BM25 implementation of Bhagavatula et al. This is likely due to the choice
of the query form that we use for \bag of words" retrieval, which is analyzed in
Section 5.3, and perhaps a better implementation of BM25 in Anserini (which
is based on Lucene).</p>
          <p>Our method appears to be as e ective and more scalable than a
clusterbased approach. For example, Bhagavatula et al.'s model requires at least 100
GB of RAM to search the 7M documents in the Open Research corpus,5 whereas
keyword search has far more modest memory requirements.</p>
          <p>In the next sections, we investigate the e ectiveness of our method by
evaluating various pretrained transformer models, as well as the e ects of class
imbalance and di erent query forms.
5.1</p>
        </sec>
      </sec>
      <sec id="sec-4-4">
        <title>In- vs. Out-Domain Pretraining</title>
        <p>Here we investigate how di erent pretraining con gurations change e ectiveness
in the target task. The results, shown in Table 3, are from ne-tuning the
pretrained models on Open Research's training set for 300k iterations with a batch
size of 128, which corresponds to approximately 1.1 epochs. In the remainder
of this paper, we call an in-domain corpus a collection whose majority of
documents are from the same domains as those in Open Research (i.e., biomedicine
and computer science), and we call an out-domain corpus a collection whose
majority of papers are not from those domains.</p>
        <p>
          The models pretrained on an in-domain corpus, i.e., BioBERT [
          <xref ref-type="bibr" rid="ref23">23</xref>
          ] (row 7)
and SciBERT [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] (rows 8{11), yield signi cant improvements in the target task
over models pretrained on a corpus of a similar size but a di erent domain (rows
3{5). Pretraining on an out-domain corpus ten times the size of the in-domain
corpus results in lower e ectiveness on the target task; compare RoBERTa [
          <xref ref-type="bibr" rid="ref25">25</xref>
          ],
row 6 vs. row 10. We conclude that, at least for the task of citation
recommendation, pretraining on a smaller in-domain corpus is more e ective than pretraining
on a larger out-domain corpus.
        </p>
        <p>
          When pretraining settings are kept the same except for the vocabulary, the
use of in-domain vocabulary gives 5{10% improvement over out-domain
vocabulary (row 8 vs. 9 and row 10 vs. 11). This make intuitive sense, and Beltagy et
al. [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ] report a similar nding in other tasks as well.
        </p>
        <p>
          The NCBI models [
          <xref ref-type="bibr" rid="ref31">31</xref>
          ] (rows 1 and 2) are pretrained on an in-domain corpus
but produce worse results than models pretrained on an out-domain corpus of
a similar size (rows 3{5). They also underperform when compared to
SciBERTBase (row 8), which is pretrained on an in-domain corpus of a similar size but
comprises full papers instead of abstracts. As also noted by Beltagy et al. [
          <xref ref-type="bibr" rid="ref3">3</xref>
          ],
this result suggests that pretraining with longer documents improves the target
task e ectiveness.
        </p>
        <p>We nd that model size appears to be even more important than document
length. Our SciBERT-Large models (rows 10 and 11) have higher e ectiveness
than the SciBERT-Base models (rows 8 and 9) despite being pretrained on a
smaller corpus of 7M paper abstracts (1.4B tokens) as opposed to 1M full-text
papers (3.2B tokens).
5.2</p>
      </sec>
      <sec id="sec-4-5">
        <title>Class Imbalance</title>
        <p>Because we only use the top 10 papers returned by BM25 as training examples,
the BERT-based models in this work are trained with more negative examples
than positive ones (94% vs. 6%). In a separate experiment, to balance these
classes, we include in the training phase pairs of query and relevant papers not
retrieved by BM25, but this results in F1@20 and MRR close to zero in both
training and development sets. We obtain a similar result when adding to the
training set negative candidates randomly sampled from the corpus.</p>
        <p>What explains these ndings? We hypothesize that although BERT is a
strong model for document ranking, it still partly relies on exact term match
to learn relevance. Thus, when we sample training documents not using an
exact term match method such as BM25, fewer terms between the query and the
candidate paper match, which makes learning relevance harder. Further studies
should investigate if this limitation applies to other tasks as well.
5.3</p>
      </sec>
      <sec id="sec-4-6">
        <title>Query Analysis</title>
        <p>In the citation recommendation task, the \query" used for initial retrieval can
take many forms, such as the title of the paper, the concatenation of title and
abstract, or keywords extracted from the text. Here we investigate how these
query forms impact the e ectiveness of a keyword-based retrieval method.</p>
        <p>
          In Table 4, we show the e ectiveness of BM25 on the Open Research
development set. For Key Terms, we follow Bhagavatula et al. [
          <xref ref-type="bibr" rid="ref4">4</xref>
          ] and use Whoosh6
to rst create an index and then extract key terms from the title and abstract
with Whoosh's key terms from text method. Despite being faster due to
having fewer query terms, the results show that this method has lower e ectiveness
than simply concatenating the title and abstract of the paper.
        </p>
        <sec id="sec-4-6-1">
          <title>6 https://whoosh.readthedocs.io/en/latest/</title>
          <p>0:11
0:164
128</p>
          <p>256
# tokens for query
384
448</p>
          <p>One of the limitations of transformer-based models (including BERT) is that
memory consumption increases quadratically with the number of tokens in the
input sequence. On modern hardware such as TPU v3s or GPU V100s, the
maximum number of tokens that we can e ciently train a BERT-Large model is
approximately 512. In our task, since the concatenation of query and candidate
tokens is typically longer than this limit, there is a trade-o between the number
of tokens we allocate to each sequence.</p>
          <p>In Figure 1, we show how e ectiveness changes as we allocate more tokens
to the query than to the candidate document while limiting the sum of the two
sequences to 512 tokens. These results are obtained with BM25 +
SciBERTBase (for faster experimental turnaround). The curve shows that query terms
are more important to the re-ranker model, as increasing query tokens from
64 to 256 increases F1@20 by 2 points. Decreasing candidate document tokens
from 256 to 64 barely changes F1@20. This result is somewhat surprising as
one expects the two sequences to have equal importance in the task of query{
document relevance estimation. Note that in all previous experiments (Table 2),
we used 256 tokens for the query and 256 for the candidate; this suggests that
our main results might be even higher had we tuned this hyperparameter as well.
Future work should investigate if this is particular to citation recommendation,
or if it also occurs in other retrieval tasks with long queries as well.
6</p>
        </sec>
      </sec>
    </sec>
    <sec id="sec-5">
      <title>Conclusions</title>
      <p>We provide an extensive evaluation of pretrained transformer models for the
scienti c literature recommendation task. We nd that in-domain pretraining
and domain-speci c vocabulary greatly improve e ectiveness. Additionally, we
present an unexpected nding: Despite the symmetry of the two inputs when
trying to estimate the relevance of a candidate article to a query article, we
nd that terms from the query article are more important than terms from
the candidate article in allocating \space" for BERT input. Future work should
investigate this observation in more detail.</p>
    </sec>
    <sec id="sec-6">
      <title>Acknowledgments</title>
      <p>This research was supported in part by the Canada First Research Excellence
Fund, the Natural Sciences and Engineering Research Council (NSERC) of
Canada, NVIDIA, and eBay. Additionally, we would like to thank Google for
computational resources in the form of Google Cloud credits.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          1.
          <string-name>
            <surname>Ammar</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Groeneveld</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bhagavatula</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Beltagy</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Crawford</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Downey</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dunkelberger</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Elgohary</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Feldman</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ha</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kinney</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kohlmeier</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lo</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Murray</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ooi</surname>
            ,
            <given-names>H.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Peters</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Power</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Skjonsberg</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wilhelm</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yuan</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          , van Zuylen,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Etzioni</surname>
          </string-name>
          ,
          <string-name>
            <surname>O.</surname>
          </string-name>
          :
          <article-title>Construction of the literature graph in Semantic Scholar</article-title>
          .
          <source>In: Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>3</volume>
          (Industry Papers). pp.
          <volume>84</volume>
          {
          <issue>91</issue>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          2.
          <string-name>
            <surname>Basu</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hirsh</surname>
          </string-name>
          , H.,
          <string-name>
            <surname>Cohen</surname>
            ,
            <given-names>W.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nevill-Manning</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          :
          <article-title>Technical paper recommendation: A study in combining multiple information sources</article-title>
          .
          <source>Journal of Arti cial Intelligence Research</source>
          <volume>14</volume>
          ,
          <volume>231</volume>
          {
          <fpage>252</fpage>
          (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          3.
          <string-name>
            <surname>Beltagy</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lo</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cohan</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          :
          <article-title>SciBERT: Pretrained contextualized embeddings for scienti c text</article-title>
          . arXiv:
          <year>1903</year>
          .
          <volume>10676</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          4.
          <string-name>
            <surname>Bhagavatula</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Feldman</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Power</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ammar</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          :
          <article-title>Content-based citation recommendation</article-title>
          . arXiv:
          <year>1802</year>
          .
          <volume>08301</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          5.
          <string-name>
            <surname>Bollacker</surname>
            ,
            <given-names>K.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lawrence</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Giles</surname>
            ,
            <given-names>C.L.:</given-names>
          </string-name>
          <article-title>A system for automatic personalized tracking of scienti c literature on the web</article-title>
          .
          <source>In: Proceedings of the Fourth ACM conference on Digital Libraries (DL '99)</source>
          . pp.
          <volume>105</volume>
          {
          <issue>113</issue>
          (
          <year>1999</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          6.
          <string-name>
            <surname>Chen</surname>
          </string-name>
          , T.T.,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Research paper recommender systems on big scholarly data</article-title>
          .
          <source>In: Paci c Rim Knowledge Acquisition Workshop</source>
          . pp.
          <volume>251</volume>
          {
          <issue>260</issue>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          7.
          <string-name>
            <surname>Devlin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chang</surname>
            ,
            <given-names>M.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Toutanova</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          : BERT:
          <article-title>Pre-training of deep bidirectional transformers for language understanding</article-title>
          .
          <source>In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies</source>
          , Volume
          <volume>1</volume>
          (Long and Short Papers). pp.
          <volume>4171</volume>
          {
          <issue>4186</issue>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          8.
          <string-name>
            <surname>Dinh</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Tamine</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Combining global and local semantic contexts for improving biomedical information retrieval</article-title>
          .
          <source>In: European Conference on Information Retrieval</source>
          . pp.
          <volume>375</volume>
          {
          <issue>386</issue>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          9.
          <string-name>
            <surname>Eto</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Extended co-citation search: Graph-based document retrieval on a cocitation network containing citation context information</article-title>
          .
          <source>Information Processing &amp; Management</source>
          <volume>56</volume>
          (
          <issue>6</issue>
          ),
          <volume>102046</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          10.
          <string-name>
            <surname>Fiorini</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Canese</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Starchenko</surname>
            ,
            <given-names>G.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kireev</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miller</surname>
            ,
            <given-names>V.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Osipov</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kholodov</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ismagilov</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mohan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ostell</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          : Best Match:
          <article-title>New relevance search for PubMed</article-title>
          .
          <source>PLoS Biology</source>
          <volume>16</volume>
          (
          <issue>8</issue>
          ),
          <year>e2005343</year>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          11.
          <string-name>
            <surname>Fiorini</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Leaman</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lipman</surname>
            ,
            <given-names>D.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          :
          <article-title>How user intelligence is improving PubMed</article-title>
          .
          <source>Nature Biotechnology</source>
          <volume>36</volume>
          (
          <issue>10</issue>
          ),
          <volume>937</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          12.
          <string-name>
            <surname>Gao</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kinoshita</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Miller</surname>
            ,
            <given-names>E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Seaborne</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cayzer</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Clark</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          :
          <article-title>Swan: A distributed knowledge infrastructure for Alzheimer disease research</article-title>
          .
          <source>Web Semantics: Science, Services and Agents on the World Wide Web</source>
          <volume>4</volume>
          (
          <issue>3</issue>
          ),
          <volume>222</volume>
          {
          <fpage>228</fpage>
          (
          <year>2006</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          13.
          <string-name>
            <surname>Ginsparg</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          :
          <article-title>First steps towards electronic research communication</article-title>
          .
          <source>Computers in Physics 8</source>
          (
          <issue>4</issue>
          ),
          <volume>390</volume>
          {
          <fpage>396</fpage>
          (
          <year>1994</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          14.
          <string-name>
            <surname>He</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pei</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kifer</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mitra</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Giles</surname>
            ,
            <given-names>C.L.</given-names>
          </string-name>
          :
          <article-title>Context-aware citation recommendation</article-title>
          .
          <source>In: Proceedings of the 19th International Conference on World Wide Web</source>
          . pp.
          <volume>421</volume>
          {
          <issue>430</issue>
          (
          <year>2010</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          15.
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kataria</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Caragea</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mitra</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Giles</surname>
            ,
            <given-names>C.L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rokach</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          :
          <article-title>Recommending citations: Translating papers into references</article-title>
          .
          <source>In: Proceedings of the 21st ACM International Conference on Information and Knowledge Management (CIKM '12)</source>
          . pp.
          <year>1910</year>
          {
          <year>1914</year>
          (
          <year>2012</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          16.
          <string-name>
            <surname>Huang</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liang</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mitra</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Giles</surname>
            ,
            <given-names>C.L.</given-names>
          </string-name>
          :
          <article-title>A neural probabilistic model for context based citation recommendation</article-title>
          .
          <source>In: Twenty-Ninth AAAI Conference on Arti cial Intelligence</source>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          17.
          <string-name>
            <surname>Jerome</surname>
            ,
            <given-names>R.N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Giuse</surname>
            ,
            <given-names>N.B.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gish</surname>
            ,
            <given-names>K.W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Sathe</surname>
            ,
            <given-names>N.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dietrich</surname>
            ,
            <given-names>M.S.:</given-names>
          </string-name>
          <article-title>Information needs of clinical teams: Analysis of questions received by the clinical informatics consult service</article-title>
          .
          <source>Bulletin of the Medical Library Association</source>
          <volume>89</volume>
          (
          <issue>2</issue>
          ),
          <volume>177</volume>
          (
          <year>2001</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          18.
          <string-name>
            <surname>Johnson</surname>
          </string-name>
          , R.,
          <string-name>
            <surname>Watkinson</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Mabe</surname>
            ,
            <given-names>M.:</given-names>
          </string-name>
          <article-title>The STM report: An overview of scienti c and scholarly publishing</article-title>
          . International Association of Scienti c,
          <source>Technical and Medical</source>
          Publishers (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          19.
          <string-name>
            <surname>Kanakia</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shen</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Eide</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>A scalable hybrid research paper recommender system for Microsoft Academic</article-title>
          .
          <source>In: The World Wide Web Conference</source>
          . pp.
          <volume>2893</volume>
          {
          <issue>2899</issue>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          20.
          <string-name>
            <surname>Kingma</surname>
            ,
            <given-names>D.P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ba</surname>
          </string-name>
          , J.:
          <article-title>Adam: A method for stochastic optimization</article-title>
          .
          <source>arXiv:1412.6980</source>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          21.
          <string-name>
            <given-names>Kodakateri</given-names>
            <surname>Pudhiyaveetil</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            ,
            <surname>Gauch</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            ,
            <surname>Luong</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            ,
            <surname>Eno</surname>
          </string-name>
          , J.:
          <article-title>Conceptual recommender system for CiteSeerX</article-title>
          .
          <source>In: Proceedings of the Third ACM Conference on Recommender Systems</source>
          . pp.
          <volume>241</volume>
          {
          <issue>244</issue>
          (
          <year>2009</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          22.
          <string-name>
            <surname>Lawrence</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bollacker</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Giles</surname>
            ,
            <given-names>C.L.</given-names>
          </string-name>
          :
          <article-title>Indexing and retrieval of scienti c literature</article-title>
          .
          <source>In: Proceedings of the 8th ACM International Conference on Information and Knowledge Management (CIKM '99)</source>
          . pp.
          <volume>139</volume>
          {
          <issue>146</issue>
          (
          <year>1999</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          23.
          <string-name>
            <surname>Lee</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yoon</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>So</surname>
            ,
            <given-names>C.H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kang</surname>
            ,
            <given-names>J.:</given-names>
          </string-name>
          <article-title>BioBERT: A pre-trained biomedical language representation model for biomedical text mining</article-title>
          . arXiv:
          <year>1901</year>
          .
          <volume>08746</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          24.
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kong</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bai</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>W.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bekele</surname>
            ,
            <given-names>T.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Xia</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          :
          <article-title>Context-based collaborative ltering for citation recommendation</article-title>
          .
          <source>IEEE Access 3</source>
          ,
          <issue>1695</issue>
          {
          <fpage>1703</fpage>
          (
          <year>2015</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          25.
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Ott</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Goyal</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Du</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Joshi</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Levy</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lewis</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Zettlemoyer</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stoyanov</surname>
          </string-name>
          , V.:
          <article-title>RoBERTa: A Robustly Optimized BERT Pretraining Approach</article-title>
          . arXiv:
          <year>1907</year>
          .
          <volume>11692</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          26.
          <string-name>
            <surname>Lu</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>He</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shan</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yan</surname>
          </string-name>
          , H.:
          <article-title>Recommending citations with translation model</article-title>
          .
          <source>In: Proceedings of the 20th ACM International Conference on Information and Knowledge Management (CIKM '11)</source>
          . pp.
          <year>2017</year>
          {
          <year>2020</year>
          (
          <year>2011</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          27.
          <string-name>
            <surname>McNee</surname>
            ,
            <given-names>S.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Albert</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cosley</surname>
            ,
            <given-names>D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gopalkrishnan</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lam</surname>
            ,
            <given-names>S.K.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Rashid</surname>
            ,
            <given-names>A.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Konstan</surname>
            ,
            <given-names>J.A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Riedl</surname>
          </string-name>
          , J.:
          <article-title>On the recommending of citations for research papers</article-title>
          .
          <source>In: Proceedings of the 2002 ACM Conference on Computer Supported Cooperative Work</source>
          . pp.
          <volume>116</volume>
          {
          <issue>125</issue>
          (
          <year>2002</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          28.
          <string-name>
            <surname>Mohan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fiorini</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kim</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          :
          <article-title>Deep learning for biomedical information retrieval: Learning textual relevance from click logs</article-title>
          .
          <source>In: BioNLP 2017</source>
          . pp.
          <volume>222</volume>
          {
          <issue>231</issue>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          29.
          <string-name>
            <given-names>Nabeel</given-names>
            <surname>Asim</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Wasim</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          , Usman Ghani Khan,
          <string-name>
            <given-names>M.</given-names>
            ,
            <surname>Mahmood</surname>
          </string-name>
          ,
          <string-name>
            <surname>W.</surname>
          </string-name>
          :
          <article-title>Improved biomedical term selection in pseudo relevance feedback</article-title>
          .
          <source>Database</source>
          <year>2018</year>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          30.
          <string-name>
            <surname>Nogueira</surname>
            ,
            <given-names>R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Cho</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          :
          <article-title>Passage re-ranking with BERT</article-title>
          . arXiv:
          <year>1901</year>
          .
          <volume>04085</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          31.
          <string-name>
            <surname>Peng</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yan</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lu</surname>
            ,
            <given-names>Z.</given-names>
          </string-name>
          :
          <article-title>Transfer learning in biomedical natural language processing: An evaluation of BERT and ELMo on ten benchmarking datasets</article-title>
          . arXiv:
          <year>1906</year>
          .
          <volume>05474</volume>
          (
          <year>2019</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          32.
          <string-name>
            <surname>Ren</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Liu</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Yu</surname>
            ,
            <given-names>X.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Khandelwal</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gu</surname>
            ,
            <given-names>Q.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wang</surname>
          </string-name>
          , L.,
          <string-name>
            <surname>Han</surname>
            ,
            <given-names>J</given-names>
          </string-name>
          .:
          <article-title>ClusCite: E ective citation recommendation by information network-based clustering</article-title>
          .
          <source>In: Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          . pp.
          <volume>821</volume>
          {
          <issue>830</issue>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          33.
          <string-name>
            <surname>Robertson</surname>
            ,
            <given-names>S.E.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Walker</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Jones</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Hancock-Beaulieu</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Gatford</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          :
          <article-title>Okapi at TREC-3</article-title>
          .
          <source>In: Proceedings of the 3rd Text REtrieval Conference (TREC-3)</source>
          . pp.
          <volume>109</volume>
          {
          <fpage>126</fpage>
          .
          <string-name>
            <surname>Gaithersburg</surname>
          </string-name>
          ,
          <string-name>
            <surname>Maryland</surname>
          </string-name>
          (
          <year>1994</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          34.
          <string-name>
            <surname>Spangler</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Wilkins</surname>
            ,
            <given-names>A.D.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Bachman</surname>
            ,
            <given-names>B.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Nagarajan</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Dayaram</surname>
            ,
            <given-names>T.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Haas</surname>
            ,
            <given-names>P.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Regenbogen</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Pickering</surname>
            ,
            <given-names>C.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Comer</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Myers</surname>
            ,
            <given-names>J.N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Stanoi</surname>
            ,
            <given-names>I.R.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Kato</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lelescu</surname>
            ,
            <given-names>A.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Labrie</surname>
            ,
            <given-names>J.J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Parikh</surname>
            ,
            <given-names>N.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lisewski</surname>
            ,
            <given-names>A.M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Donehower</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Chen</surname>
            ,
            <given-names>Y.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lichtarge</surname>
            ,
            <given-names>O.</given-names>
          </string-name>
          :
          <article-title>Automated hypothesis generation based on mining scienti c literature</article-title>
          .
          <source>In: Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          . pp.
          <year>1877</year>
          {
          <year>1886</year>
          (
          <year>2014</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          35.
          <string-name>
            <surname>Sybrandt</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Shtutman</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Safro</surname>
            ,
            <given-names>I.</given-names>
          </string-name>
          :
          <article-title>Moliere: Automatic biomedical hypothesis generation system</article-title>
          .
          <source>In: Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining</source>
          . pp.
          <volume>1633</volume>
          {
          <issue>1642</issue>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          36.
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          :
          <article-title>Anserini: Enabling the use of Lucene for information retrieval research</article-title>
          .
          <source>In: Proceedings of the 40th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR</source>
          <year>2017</year>
          ). pp.
          <volume>1253</volume>
          {
          <issue>1256</issue>
          (
          <year>2017</year>
          )
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          37.
          <string-name>
            <surname>Yang</surname>
            ,
            <given-names>P.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Fang</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          ,
          <string-name>
            <surname>Lin</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          : Anserini:
          <article-title>Reproducible ranking baselines using Lucene</article-title>
          .
          <source>Journal of Data and Information Quality</source>
          <volume>10</volume>
          (
          <issue>4</issue>
          ),
          <source>Article</source>
          <volume>16</volume>
          (
          <year>2018</year>
          )
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>