<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Entity Retrieval Docker Image for OSIRRC at SIGIR 2019</article-title>
      </title-group>
      <contrib-group>
        <aff id="aff0">
          <label>0</label>
          <institution>Negar Arabzadeh Ryerson University Toronto</institution>
          ,
          <addr-line>Ontario</addr-line>
        </aff>
      </contrib-group>
      <fpage>21</fpage>
      <lpage>25</lpage>
      <abstract>
        <p>With emerging of structured data, retrieving entities instead of documents becomes more prevalent in order to satisfy the information need related to a query. Therefore, several high-performance entity retrieval methods have been introduced to the Information Retrieval (IR) community in recent years. Replicating and reproducing the standard entity retrieval methods are considered as challenging tasks in the IR community. Open-Source IR Replicability Challenge (OSIRRC) has addressed this problem by introducing a unified framework for dockerizing a variety of retrieval tasks. In this paper, a Docker image is built for six diferent entity retrieval models including, LM, MLM-tc, MLM-all, PRMS, SDM, FSDM. Also, Entity Linking incorporated Retrieval(ELR) extension, has been implemented that can be applied on top of all the mentioned models. The entity retrieval docker can retrieve relevant entities for any given topic. Image Source: https://github.com/osirrc/entityretrieval-docker Docker Hub: https://hub.docker.com/r/osirrc2019/entityretrieval</p>
      </abstract>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>OVERVIEW</title>
      <p>
        In the past two decades, search engines have been dealing with
unorganized and unclassified data i.e., unstructured data until the
emergence of semantic search. In order to satisfy the information
need behind a query using structured data, retrieving
machinerecognizable "entities" has proven to be a suitable complementary
for document retrieval for multiple reasons. For instance, Returning
a document in response to a query that is looking for an entity might
not be the best option because users have to look into the document
to find their desired information need. That could be one of the
main reason why retrieval operations are getting more and more
entity-centric, particularly on the web. Document retrieval difers
from entity retrieval in a couple of senses. In document retrieval,
entities are usually used for query expansion or retrieval features in
order to improve learning-to-rank frameworks and consecutively
to enhance document retrieval performance. On the other hand,
entity retrieval is defined as searching for an entity in a knowledge
base where entities are first class citizens[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. In other words, our
goal is to retrieve the most relevant entities from a knowledge base
for a given term-based query.
      </p>
      <p>In this docker image, the following standard entity retrieval
models have been implemented:</p>
      <p>
        • LM [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ]
• MLM-tc [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ]
• MLM-all [
        <xref ref-type="bibr" rid="ref8">8</xref>
        ]
• PRMS [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ]
• SDM [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ]
• FSDM [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]
      </p>
      <p>
        Furthermore, an extension of the Markov Random Field (MRF)
model framework for incorporating entity annotations into the
retrieval model, which is called Entity linking incorporated
Retrieval(ELR) has been leveraged on top of mentioned retrieval model.
Applying ELR on the state-of-the-art entity retrieval models results
in having the following ELR-integrated entity retrieval models [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] :
• LMelr
• MLM − tcelr
• MLM − allelr
• PRMSelr
• SDMelr
• F SDMelr
      </p>
      <p>DBpedia version 3.9 has been used as the knowledge base in
the entity retrieval tasks in this Docker image. A term-based index
and an entity-based index had created from the knowledge base
by utilizing Lucene. Within the former index, entities URI objects
are resolved to terms and the default Lucene stop words have been
removed from them. Meanwhile, within the latter, URI objects are
preserved and literal objects are ignored. In total, entity-based index
contains 3,984,580 entities. While the term-based index is used in
the standard entity retrieval model, both term based and entity
based indices are used in ELR entity retrieval models.</p>
      <p>
        One of the critical components of the ELR approaches is entity
annotations of the queries. TAGME, which is an open source entity
linker, has been adopted to annotate entities in queries with the
default threshold 0.1. Hasibi et al.[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ] have shown that the ELR
approach is robust to annotation threshold.
      </p>
      <p>In summary, given a query Q and a retrieval model M, the
entity retrieval operates according to Figure 1, where D is the
representation of entities. In section 2, there will be a more in-depth
elaboration on standard entity retrieval methods in addition to the
combination of them with ELR extension. Section 3 provides more
details on the technical design aspect of the docker image. Section
4 describes our motivation and experience in participating in the
OSIRRC 2019 challenge. The last section, i.e., section 5, gives an
insight on further work that has to be carried out in this area and
conclude the paper.
2</p>
    </sec>
    <sec id="sec-2">
      <title>RETRIEVAL MODELS</title>
      <p>
        As was mentioned in the Overview section, several standard
retrieval models including LM, MLM-tc, MLM-all, PRMS, SDM and
FSDM have been implemented in this docker. In addition, ELR
extension can be applied on top of them, which results in having
twelve diferent retrieval models collectively. There will be two
different representation of entities; the term-based representation and
the entity-based representation. For the standard retrieval models,
term-based representation of DBpedia collection (using term-based
index) is used and when it comes to ELR extension. On the other
hand, both term-based and entity-based representation of entities
in DBpedia are used (term-based and URI-based indices). The
termbased and entity-based representations are compared in Figure 2
[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
      </p>
      <p>Retrieval models can be categorized into Language
modelingbased models and Sequential Dependence models and ELR models.
2.1</p>
    </sec>
    <sec id="sec-3">
      <title>Language Modeling-based methods</title>
      <p>
        Language modeling-based models consider dependencies among
query terms. LM [
        <xref ref-type="bibr" rid="ref9">9</xref>
        ], MLM-tc[
        <xref ref-type="bibr" rid="ref7">7</xref>
        ], MLM-all[
        <xref ref-type="bibr" rid="ref8">8</xref>
        ] and PMRS[
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] are
all language modeling based methods. LM only applies on the
content field. However, MLM-tc run against name field as well
as content field . The fields content and name have weights of
0.8 and 0.2 respectively. MLM-all and PRMS use top 10 fields . The
former is a Mixture of Language models with top 10 fields with
uniform weights but the latter’s retrieval model is a probabilistic
one designed for semi-structured data. More details on each of the
mentioned language models can be found in the related papers.
2.2
      </p>
    </sec>
    <sec id="sec-4">
      <title>Sequential Dependence-based methods</title>
      <p>Sequential Dependence Models (SDM) are popular Markov Random
Field based retrieval models. Given a document D and a query Q, the
conditional probability of P(D|Q) is estimated based on Markov
Random Field as in equation (1) where D is term-based representation
of entities.</p>
      <p>rank
P (D |Q) =</p>
      <p>
        In equation (1), C(G) is set of cliques in graph G . The nodes of the
graph G consist of query terms and documents and the edges among
nodes illustrates the dependencies among nodes. λc is weight of
the feature function f(c) . More details can be found in the original
paper [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ].
      </p>
      <p>Considering dependencies among query terms results in having
equation (2) based on Markov Random Field (equation (1)) as SDM
ranking function with respect to λT + λO +λU = 1 :
(1)
(2)</p>
      <p>P (D |Q)
rank
= λT
Õ
2.2.1 Fielded Sequential Dependence Models (FSDM) .</p>
      <p>
        Fielded Sequential Dependence Models (FSDM) considers
document structure by computing linear interpolation of probability of
each documents’ fields.Thus, feature functions are also calculated
based on field representation of documents. In other words, in FSDM
model [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ] equation (2) comes with diferent feature functions as
diferent language models are built for each field.
2.3
      </p>
    </sec>
    <sec id="sec-5">
      <title>ELR models</title>
      <p>
        Incorporating Entity Linking into entity Retrieval leads to improve
entity retrieval performance [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. Linking entities by TAGME results
in having confidence score s(e) for each entity e. While considering
sequential dependency in MRF-based models, annotated queries are
assumed to be independent of each other and query terms. Applying
ELR extension on the previous models results in the equation(3)
as ranking function where |Q | is the query length and s(e) is the
entity linking confidence score of entity e annotated by TAGME.
Equation(3) is elaborated more in [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] with its feature functions.
Free parameters constraints of λT + λO + λU + λE = 1 is true for
equation 3. For LM, MLM-tc and MLM-all λO and λU are set to
zero since they are unigram based models. All the feature functions
are defined in [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
One of the major issues when dealing with replicability problem,
is that the system should be delivered in a lightweight package[
        <xref ref-type="bibr" rid="ref1">1</xref>
        ].
Dockers has this ability since a relatively inexpensive container
can be created from each docker image. a jig was introduced in
OSIRRC2019 workshop that makes the co-implementing and
codesigning available. The jig which is open source and available on
GitHub1 plays a semi-tool role which can maintain computational
relationship among Dockers and retrieval tasks.
      </p>
      <p>
        The entity retrieval Docker image is consisted of init, index
and search hooks invokes by Python3 as its interpreter. The jig
triggers the init hook first and then index and search respectively
in the Docker image. Finally , if there golden standard for the topics
are available, it will evaluate the results. Since we can get data into
and out of the container built from the Docker image [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ], we can get
the retrieval results in the jig output directory. further explanation
to run the entity retrieval Docker image is available on the entity
retrieval GitHub repository.
      </p>
      <p>This section describes diferent components of the docker image
and supported hooks and extra options which can be passed to the
jig for the entity retrieval Docker.
3.1</p>
    </sec>
    <sec id="sec-6">
      <title>Dockerfile</title>
      <p>The latest oficial version of Ubuntu 2 is installed in the Docker
with all the required commands. In addition, compatible versions
of other requirements such as java8, Apache Ant, Apache Ivy, g+,
and so on are installed on the Docker. Making all the components
compatible with each other was a quite challenging issue. Since
installing all the requirements is a time-consuming step, a docker
image is prepared with all the basic requirements and pushed it to
Docker Hub3. Hence, this prepared image is used as our Dockerfile
base image in order to decrease the building time of the Docker.
This sets the stage for COPYing the init,index and search hooks
which should be executable files.
3.2</p>
    </sec>
    <sec id="sec-7">
      <title>Supported Collections</title>
      <p>DBpedia version 3.9 4 has been used as the corpus of entity retrieval
Docker image. In order to reduce the run time cost of the docker,
the original index is used. Both term-based indexed and URI-based
indexed of the collection will be downloaded once in the preparation
step of the jig. However, to make the docker run using "the jig", a
dummy collection is needed to pass.
3.3</p>
    </sec>
    <sec id="sec-8">
      <title>Supported Hooks</title>
      <p>This section elaborates on the role of each hook separately. init and
index hooks are triggered in the jig preparation step and search
hook script will run in the jig search step.</p>
      <p>3.3.1 init.</p>
      <p>The actual implementation of the retrieval models is cloned in
this hook from the GitHub repository5. The required compatible
packages are installed. Running this hook may take a while
because of downloading,building and installing PyLucene which is
time-consuming. Once the installation are done, two indexed
collection which are DBpedia term-based index and URI-based index
are downloaded (∼ 18GB) and extracted.</p>
      <p>3.3.2 index.</p>
      <p>The indexed DBpedia collection is already downloaded in the init
hook. Hence, nothing is happening in this hook in this docker.</p>
      <p>3.3.3 search.</p>
      <p>
        When the image is prepared, retrieval models can run with respect
to their relevant customized parameters in the search hook. The
search hook runs the main implementation of models6 which were
provided by Hasibi et al.[
        <xref ref-type="bibr" rid="ref2">2</xref>
        ]. the main code was cloned in init hook
in the Docker. Table 1 demonstrates all the parameters that can be
set for each of the retrieval models. Given the query , depending
on the retrieval model, the query would be annotated or not, and
then the retrieval takes place based on the set parameters. Then,
2https://hub.docker.com/_/ubuntu
3https://hub.docker.com/r/narabzad/elr_prepared_os
4https://wiki.dbpedia.org/services-resources/datasets/data-set-39/downloads-39
5https://github.com/Narabzad/elr_files
6https://github.com/hasibi/EntityLinkingRetrieval-ELR/
the ranked list of retrieved entities for each query will be saved in
the output repository.
      </p>
      <p>All the queries in the jig e.g, topics of Robust04, ClueWeb09,
ClueWeb12, Gov2, Core17, Core18 and etc. are supported in this
Docker. Any other queries is acceptable in this docker as long as
each query is represented in the following format "query number
or name:query terms" in each line. An instance of a topic file would
be like:
wt09-1:obama family tree
wt09-2:french lick resort and casino
wt09-3:getting organized
...</p>
      <p>If relevant entities (qrel) are available for the associated topics,
the jig will utilize Trec Eval 7 to evaluate the retrieval performance
by diferent metrics such as MAP. If there is no ground truth ranked
list of entities for the topic, a dummy qrel file is needed to pass to
the jig in order to make the Docker run by the jig.
4</p>
    </sec>
    <sec id="sec-9">
      <title>OSIRRC EXPERIENCE</title>
      <p>
        The crucial role of repeatability, replicability, and reproducibility
cannot be neglected in any research domain; especially when it
comes to practical experiments. Deciding to participate in this
challenge was easy because either it is repeating your computation,
replicating another researchers’ experiments or reproducing other
team’s research with a totally diferent setup, there will be an
endeavored struggle. This docker tackles all these 3 challenges for
entity retrieval. Not only the docker is built by a non-author of the
main paper (replicability) [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], but also this work is not limited to
specific topics. In other words, the entity retrieval docker is
modiifed in a way that any topics can be used to retrieve the relevant
7https://github.com/usnistgov/trec_eval
entities (reproducibility) and if the relevant entities i.e., the golden
standard, is available for the topics, the model can be evaluated
as well. However, the supported collection is still limited to the
indexed DBpedia version 3.9.
      </p>
      <p>Dockerizing the entity retrieval models was a challenging task.
furthermore, standardizing the Docker with the jig increased its
complexity. One of the main issues regarding this Docker was
the compatibility of diferent components e.g, Python, PyLucene,
Java, Apache Ant, etc. It was a time-consuming task to find all the
compatible version of all those components and this is one of the
critical benefits of this task. Utilizing the entity Docker, Researchers
do not have to spend lots of time on combining and connecting
diferent packages, libraries and components anymore to run the
mentioned entity retrieval models.</p>
      <p>Another issue is that topics appear in diferent formats. we must
be able to work with every topic format available in the jig.
Therefore a standard topic format is defined in section 3.3.3 so that any
topic can be used in this Docker.</p>
      <p>
        For the methods with ELR extension, the annotation step has
to be added to the code that was implemented by Hasibi et al. [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ].
Utilizing TAGME tool results in linking entities to queries.
5
      </p>
    </sec>
    <sec id="sec-10">
      <title>FUTURE WORKS AND CONCLUDING</title>
    </sec>
    <sec id="sec-11">
      <title>REMARKS</title>
      <p>The more reproducible and replicable research papers are, the more
baselines will be accessible for researches to compare their results
with. This means, by increasing repeatability, reproducibility,and
replicability researchers can can spend less time on implementing
other researchers work. Therefore, they will have more time
spending on their own research and not on implementing the baselines.
Consecutively, studies would make progress faster. Hence, more
works needed to be done in this specific area.</p>
      <p>
        According to entity retrieval docker image, this work can be
extended by supporting more collection such as DBpedia-entity v2
[
        <xref ref-type="bibr" rid="ref4">4</xref>
        ] in addition to the current one. Nordlys [
        <xref ref-type="bibr" rid="ref3">3</xref>
        ] implements some of
the models with the updated collection. So in the future, a Docker
could be created for Nordlys, which provides better support for
indexing.
      </p>
      <p>Furthermore, entity retrieval models can be added to the Docker.
In terms of entity retrieval applications, it can be used to expand
queries in order to improve document retrieval performance.</p>
      <p>To sum up our work, Docker image is wrapped around the jig
introduced for Open-Source IR Replicability Challenge 2019. This
platform provides a unified framework for diferent retrieval task.
Entity retrieval Docker image contains implementation of six
diferent entity retrieval model. ELR extension also can be applied on any
of the models. All models can be customized with desired
parameters and there is no limit in the supported topics. This docker image
is implemented based on very lightweight Linux-centric design to
tackle repeatability, reproducibility and replicability problem in the
IR domain.</p>
    </sec>
    <sec id="sec-12">
      <title>ACKNOWLEDGEMENT</title>
      <p>Thanks to Faegheh Hasibi for her valuable suggestions during the
implementation of the Docker image and preparation of this paper.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>Ryan</given-names>
            <surname>Clancy</surname>
          </string-name>
          , Nicola Ferro, Claudia Hauf, Jimmy Lin, Tetsuya
          <string-name>
            <surname>Sakai</surname>
          </string-name>
          , and Ze Zhong Wu.
          <year>2019</year>
          .
          <article-title>The SIGIR 2019 Open-Source IR Replicability Challenge (OSIRRC</article-title>
          <year>2019</year>
          )
          <article-title>+</article-title>
          . https://doi.org/10.1145/3331184.3331647
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>Faegheh</given-names>
            <surname>Hasibi</surname>
          </string-name>
          , Krisztian Balog, and Svein Erik Bratsberg.
          <year>2016</year>
          .
          <article-title>Exploiting Entity Linking in Queries for Entity Retrieval</article-title>
          .
          <source>In Proceedings of the 2016 ACM on International Conference on the Theory of Information Retrieval</source>
          ,
          <string-name>
            <surname>ICTIR</surname>
          </string-name>
          <year>2016</year>
          , Newark, DE, USA,
          <source>September 12- 6</source>
          ,
          <year>2016</year>
          .
          <fpage>209</fpage>
          -
          <lpage>218</lpage>
          . https://doi.org/10.1145/2970398.2970406
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>Faegheh</given-names>
            <surname>Hasibi</surname>
          </string-name>
          , Krisztian Balog, Darío Garigliotti,
          <string-name>
            <given-names>and Shuo</given-names>
            <surname>Zhang</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>Nordlys: A Toolkit for Entity-Oriented and Semantic Search</article-title>
          .
          <source>In Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          , Shinjuku, Tokyo, Japan,
          <source>August</source>
          <volume>7</volume>
          -
          <issue>11</issue>
          ,
          <year>2017</year>
          .
          <fpage>1289</fpage>
          -
          <lpage>1292</lpage>
          . https://doi.org/ 10.1145/3077136.3084149
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>Faegheh</given-names>
            <surname>Hasibi</surname>
          </string-name>
          , Fedor Nikolaev, Chenyan Xiong, Krisztian Balog, Svein Erik Bratsberg, Alexander Kotov, and
          <string-name>
            <given-names>Jamie</given-names>
            <surname>Callan</surname>
          </string-name>
          .
          <year>2017</year>
          .
          <article-title>DBpedia-Entity v2: A Test Collection for Entity Search</article-title>
          .
          <source>In Proceedings of the 40th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          , Shinjuku, Tokyo, Japan,
          <source>August</source>
          <volume>7</volume>
          -
          <issue>11</issue>
          ,
          <year>2017</year>
          .
          <fpage>1265</fpage>
          -
          <lpage>1268</lpage>
          . https://doi.org/10.1145/3077136.3080751
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>Jinyoung</given-names>
            <surname>Kim</surname>
          </string-name>
          , Xiaobing Xue, and
          <string-name>
            <given-names>W. Bruce</given-names>
            <surname>Croft</surname>
          </string-name>
          .
          <year>2009</year>
          .
          <article-title>A Probabilistic Retrieval Model for Semistructured Data</article-title>
          .
          <source>In Advances in Information Retrieval, 31th European Conference on IR Research</source>
          , ECIR
          <year>2009</year>
          , Toulouse, France, April 6-
          <issue>9</issue>
          ,
          <year>2009</year>
          . Proceedings.
          <volume>228</volume>
          -
          <fpage>239</fpage>
          . https://doi.org/10.1007/978-3-
          <fpage>642</fpage>
          -00958-7_
          <fpage>22</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>Donald</given-names>
            <surname>Metzler</surname>
          </string-name>
          and
          <string-name>
            <given-names>W. Bruce</given-names>
            <surname>Croft</surname>
          </string-name>
          .
          <year>2005</year>
          .
          <article-title>A Markov random field model for term dependencies</article-title>
          .
          <source>In SIGIR 2005: Proceedings of the 28th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          , Salvador, Brazil,
          <source>August 15-19</source>
          ,
          <year>2005</year>
          .
          <fpage>472</fpage>
          -
          <lpage>479</lpage>
          . https://doi.org/10.1145/1076034.1076115
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>Robert</given-names>
            <surname>Neumayer</surname>
          </string-name>
          , Krisztian Balog, and
          <string-name>
            <given-names>Kjetil</given-names>
            <surname>Nørvåg</surname>
          </string-name>
          .
          <year>2012</year>
          .
          <article-title>When Simple is (more than) Good Enough: Efective Semantic Search with (almost) no Semantics</article-title>
          .
          <source>In Advances in Information Retrieval - 34th European Conference on IR Research</source>
          , ECIR
          <year>2012</year>
          , Barcelona, Spain, April 1-
          <issue>5</issue>
          ,
          <year>2012</year>
          . Proceedings.
          <volume>540</volume>
          -
          <fpage>543</fpage>
          . https://doi.org/ 10.1007/978-3-
          <fpage>642</fpage>
          -28997-2_
          <fpage>59</fpage>
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>Paul</given-names>
            <surname>Ogilvie</surname>
          </string-name>
          and
          <string-name>
            <given-names>James P.</given-names>
            <surname>Callan</surname>
          </string-name>
          .
          <year>2003</year>
          .
          <article-title>Combining document representations for known-item search</article-title>
          .
          <source>In SIGIR 2003: Proceedings of the 26th Annual International ACM SIGIR Conference on Research and Development in Information Retrieval, July 28 - August 1</source>
          ,
          <year>2003</year>
          , Toronto, Canada.
          <fpage>143</fpage>
          -
          <lpage>150</lpage>
          . https://doi.org/10.1145/860435. 860463
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>ChengXiang</given-names>
            <surname>Zhai</surname>
          </string-name>
          .
          <year>2008</year>
          .
          <article-title>Statistical Language Models for Information Retrieval: A Critical Review</article-title>
          .
          <source>Foundations and Trends in Information Retrieval 2</source>
          ,
          <issue>3</issue>
          (
          <year>2008</year>
          ),
          <fpage>137</fpage>
          -
          <lpage>213</lpage>
          . https://doi.org/10.1561/1500000008
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>Nikita</surname>
            <given-names>Zhiltsov</given-names>
          </string-name>
          , Alexander Kotov, and
          <string-name>
            <given-names>Fedor</given-names>
            <surname>Nikolaev</surname>
          </string-name>
          .
          <year>2015</year>
          .
          <article-title>Fielded Sequential Dependence Model for Ad-Hoc Entity Retrieval in the Web of Data</article-title>
          .
          <source>In Proceedings of the 38th International ACM SIGIR Conference on Research and Development in Information Retrieval, Santiago, Chile, August</source>
          <volume>9</volume>
          -
          <issue>13</issue>
          ,
          <year>2015</year>
          .
          <fpage>253</fpage>
          -
          <lpage>262</lpage>
          . https://doi.org/ 10.1145/2766462.2767756
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>