<!DOCTYPE article PUBLIC "-//NLM//DTD JATS (Z39.96) Journal Archiving and Interchange DTD v1.0 20120330//EN" "JATS-archivearticle1.dtd">
<article xmlns:xlink="http://www.w3.org/1999/xlink">
  <front>
    <journal-meta />
    <article-meta>
      <title-group>
        <article-title>Integrating Query Interpretation Components into the Information Retrieval Experiment Platform</article-title>
      </title-group>
      <contrib-group>
        <contrib contrib-type="author">
          <string-name>Marcel Gohsen</string-name>
        </contrib>
        <contrib contrib-type="author">
          <string-name>Benno Stein</string-name>
        </contrib>
        <contrib contrib-type="editor">
          <string-name>Bauhaus-Universität Weimar</string-name>
        </contrib>
      </contrib-group>
      <abstract>
        <p>We describe our contribution of query entity linking and query interpretation software as components of the Information Retrieval Experiment Platform (TIREx) to the 1st International Workshop on Open Web Search (WOWS'24). Query interpretation is the task of determining all plausible user intents behind a search query and can be used to diversify the search results of a retrieval system. As TIREx components, the query entity linking method has identified a total of 89,289 entity candidates in 2,544 queries from 31 standard information retrieval datasets. In addition, a total of 2,304 interpretations of 1,225 search queries from 18 keyword query datasets have been found, which means that on average there is more than one plausible interpretation per search query. This is an indication that most search queries are ambiguous.</p>
      </abstract>
      <kwd-group>
        <kwd>eol&gt;Query understanding</kwd>
        <kwd>Query interpretation</kwd>
        <kwd>Query entity linking</kwd>
        <kwd>Query segmentation</kwd>
      </kwd-group>
    </article-meta>
  </front>
  <body>
    <sec id="sec-1">
      <title>1. Introduction</title>
      <p>
        The Information Retrieval Experiment Platform (TIREx) [
        <xref ref-type="bibr" rid="ref1">1</xref>
        ] integrates ir_datasets [
        <xref ref-type="bibr" rid="ref2">2</xref>
        ], ir_measures
[
        <xref ref-type="bibr" rid="ref3">3</xref>
        ], PyTerrier [
        <xref ref-type="bibr" rid="ref4">4</xref>
        ], and the TIRA Integrated Research Architecture [
        <xref ref-type="bibr" rid="ref5">5</xref>
        ] in order to provide an open-source
platform to expriment with information retrieval datasets and evaluate retrieval systems in a reproducible
fashion. A key aspect of TIREx is that information retrieval collections are static, and therefore query
processors only need to be executed once per dataset, so that downstream experiments can utilize the
cached and publicly available results instead of executing the processors from scratch.
      </p>
      <p>
        To streamline further research towards query understanding, we contribute query entity linking
and query interpretation components from Kasturia et al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] to the TIREx platform. The foundation
for interpreting search queries is query segmentation, a method for grouping keywords of a search
query into phrases with the aim of maximizing retrieval eficiency when these phrases are matched
with the search results. Segmentations of a query form “skeletons” for query interpretations, in which
segments are replaced by linked entities when a segment refers to an entity. Each diferent entity-linked
segmentation (i.e., interpretation) represents a diferent user intent. For example, interpretations of the
query “new york times square dance” can be either ⟨New_York_City|Times_Square|dance⟩ which
refers to a dance event on the Times Square in New York City or ⟨New_York_Times|Square_Dance⟩
which references an article in the New York Times about square dancing.
      </p>
      <p>
        As demonstrated by the example above, search engine queries can be ambiguous. Automated
approaches to interpreting search queries can help a search engine understand a user’s intent or
diversify search results based on all possible interpretations. Moreover, prior studies have shown that
extending queries with (linked) named entities can increase the efectiveness of sparse retrieval [
        <xref ref-type="bibr" rid="ref7">7</xref>
        ],
entity retrieval [
        <xref ref-type="bibr" rid="ref8 ref9">8, 9</xref>
        ] and semantic search [
        <xref ref-type="bibr" rid="ref10">10</xref>
        ]. Integrating query entity linking and query interpretation
components into TIREx facilitates further research in these directions.
      </p>
      <p>In this paper, we describe the query entity linking and query interpretation components and report
on entity and interpretation statistics of queries from standard information retrieval datasets. As part
of these exemplary analytics, we find that on average queries from almost all common information
retrieval datasets have more than one plausible interpretation. We also observe that the number of
interpretations of a query does not correlate with the number of relevant documents. Consequently,
this opens a direction for future work to determine whether query interpretations as an intermediate
step in retrieval would increase the number of relevant search results when a query is ambiguous (i.e.,
the number of plausible interpretations is greater than one).</p>
    </sec>
    <sec id="sec-2">
      <title>2. Entity-based Query Interpretation</title>
      <p>
        The query interpretation approach of Kasturia et al. [
        <xref ref-type="bibr" rid="ref6">6</xref>
        ] consists of three main phases: entity linking,
query segmentation and a combination phase, in which the linked entities and query segmentations
are combined into query interpretations. The result of the query interpretation approach is a list of
entity-linked segmentations ranked by a relevance score.
      </p>
      <sec id="sec-2-1">
        <title>2.1. Query Entity Linking</title>
        <p>The entity linking approach tries to find entities for all the (2) potential segments of an -term query.
The entity linking module is based on titles of Wikipedia articles, redirects, and disambiguation pages.
Each Wikipedia article represent an entity and its title, all redirect candidates, and disambiguation
names from Wikipedia serve as plausible query segments that refer to this entity. The about 13 million
distinct key–value pairs (keys are potential query segments and values are lists of entities that can be
referred to by this segment) are stored in a RocksDB table1 for fast exact-match access and in a Lucene
index2 to quickly find imperfect matches.</p>
        <p>
          Entities from perfect and imperfect matches are then ranked by commonness scores (i.e., the likelihood
of an entity-mention link). To compute the commonness of a mention-entity pair, a Wikipedia dump in
combination with the computation methodology from Ferragina and Scaiella [
          <xref ref-type="bibr" rid="ref11">11</xref>
          ] is used.
        </p>
        <p>
          Table 1 displays the identified and linked entities in two queries from TREC Web Track 2009 [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ] and
2012 [
          <xref ref-type="bibr" rid="ref13">13</xref>
          ] datasets. The top four entities linked to segments in the query “obama family tree” are highly
relevant and are assessed as such with commonness scores greater than 0.29. Less relevant entities
like the TV series or the music album “Family Tree” have received scores of 0.07 and less. In contrast,
scores of entities in the query “pork tenderloin” seem to be less informative. While Tenderloin in San
Francisco can still be part of a meaningful interpretation (i.e., the user is looking for pork in Tenderloin,
San Francisco), the other entities have a rather weak connection to pork. However, these entities have
been awarded a comparatively low score of 0.12 or less.
        </p>
      </sec>
      <sec id="sec-2-2">
        <title>2.2. Query Segmentation</title>
        <p>
          Query segmentation methods aim to rank the possible 2 valid segmentations of an -term query
according to retrieval efectiveness when the segments are treated as phrases to be matched in the
search results. The query interpretation approach employs query segmentation approaches from Hagen
et al. [
          <xref ref-type="bibr" rid="ref14 ref15">14, 15</xref>
          ] due to the simplicity of the approaches and the associated lower runtime. These query
segmentation approaches rank the possible segmentations of a query by summing up pre-computed
segment weights stored in a hash table for quick access. The segment weights are occurence frequencies
from the Google n-gram corpus3 in case a segment does not represent a title of a Wikipedia article. In
case this segment is the title of a Wikipedia article, the segment weight is 1 + occurrence frequency of
the most frequent word-2-gram in that segment.
        </p>
        <p>
          Most of the 2 segmentations of an -term query do not yield plausible interpretations. Kasturia et al.
[
          <xref ref-type="bibr" rid="ref6">6</xref>
          ] have shown that often only the segmentation with the highest score is used as an interpretation
skeleton and that segmentations with lower scores, which show large diferences to segmentations
with higher scores, are rarely used as skeletons. Therefore, respective filter heuristics are applied to
not forward all segmentations to the combination phase. A first filter removes segmentations whose
        </p>
        <sec id="sec-2-2-1">
          <title>1https://rocksdb.org/ 2https://lucene.apache.org/ 3https://catalog.ldc.upenn.edu/LDC2006T13</title>
          <p>highest weighting segment is contained in a higher-ranked segmentation. A second filter removes
segmentations for which the score ratio to the lowest kept higher-ranked segmentation falls below a
threshold of 0.66.</p>
        </sec>
      </sec>
      <sec id="sec-2-3">
        <title>2.3. Query Interpretation</title>
        <p>To build query interpretations for a query, the approach combines segmentations with linked entities.
To “fill” the segmentation skeletons forwarded by the segmentation phase, the approach collects the
entities ranked by commonness for each segment (discarding entities with a commonness of 0) and
adds the option of not linking a segment to an entity, but keeping it as a phrase as a fallback. The
potential interpretations can then be derived by a Cartesian product of the not-0-common entities and
the unlinked respective segments.</p>
        <p>
          In the interpretation ranking, three weights from the entity linking literature are combined: (1) the
above described commonness CMN, (2) the likelihood of two entities to occur together
(relatedness REL), and (3) the likelihood of an entity to occur with the unlinked segments (context CXT). The
relatedness and context weights computations are calculated by using Wikipedia-based joint word–
entity embeddings4 provided by Yamada et al. [
          <xref ref-type="bibr" rid="ref16">16</xref>
          ]. The by the authors suggested configuration have
been used: average cosine similarity of an entity’s embedding vector with the other entities in an
interpretation (relatedness) or with the unlinked segments in an interpretation (context).
        </p>
        <p>
          An interpretation ’s score is the averaged weighted sum of the commonness, relatedness, and context
scores of the entities  ∈  with the weights  =  =  = 1 suggested by Kasturia et al. [
          <xref ref-type="bibr" rid="ref6">6</xref>
          ]:
score() =
1
        </p>
        <p>∑︁ ( · CMN() +  · REL() +  · CXT()) ,
|{ ∈ }| · ∈</p>
        <p>Table 2 presents the found query interpretations for the example queries from Table 1 and an additional
query from TREC Web Track 2012. For the query “obama family tree” the most plausible interpretations
have been identified, which is that a user is looking up the family tree of Barack Obama. These
interpretations are on rank 1 and rank 2 and can be seen as equivalent because the concept of a family
tree is semantically equivalent to the linked entity Family_Tree. The interpretation on rank 3 assumes
that the segment “obama” is a concept and not an entity which is wrong, but the assigned relevance is
desirably low. For the query “pork tenderloin” the only found interpretation is the dish pork tenderloin.
Desirabel (but less likely) would also be the interpretation ⟨pork | Tenderloin,_San_Francisco⟩ as</p>
        <sec id="sec-2-3-1">
          <title>4https://wikipedia2vec.github.io/wikipedia2vec/</title>
          <p>someone might look for meat vendors in Tenderloin in San Francisco. However, the system did not
provide this plausible interpretation. Another interesting example is the query “last supper painting”.
The system identifies three plausible interpretations which translate to a user who is interested in (1) the
painting process of the painting “Last Supper” by Leonardo da Vinci, (2) any painting that depicts the
last supper, and (3) information about “Last Supper” by Leonardo da Vinci. All three interpretations are
highly likely which is reflected in the assigned relevance scores.</p>
        </sec>
      </sec>
      <sec id="sec-2-4">
        <title>2.4. Integration into TIREx</title>
        <p>Both, the query entity linking and the query interpretation software, are Java-based implementations
that require external data (e.g., RocksDB and Lucene Index for entity linking) to function as intended.
To integrate these components into TIREx and ensure reproducible execution across all systems, we
dockerize the query interpretation and entity linking software and bundle the image with all necessary
external data. Both Docker images are available in our public container registry5 to use outside
TIREx. The use in combination with TIREx is documented in examplary Jupyter notebooks in the code
repositories of query entity linking6 and query interpretation7 components.</p>
      </sec>
    </sec>
    <sec id="sec-3">
      <title>3. Query Analytics</title>
      <p>To gain insights into the identified entities and interpretations, we perform a statistical analysis of the
entity and interpretation frequencies of queries from common information retrieval datasets available
in TIREx.</p>
      <sec id="sec-3-1">
        <title>3.1. Datasets</title>
        <p>
          TIREx accesses datasets through the Python package of ir_datasets [
          <xref ref-type="bibr" rid="ref2">2</xref>
          ]. Therefore, all the processed
datasets are well documented in the online catalog8 of ir_datasets. Most of the 31 considered
datasets originate from shared tasks like TREC or Touché. We distinguish these datasets by query type
which are keyword queries (e.g., “french lick resort and casion” from TREC Web Track 2009 [
          <xref ref-type="bibr" rid="ref12">12</xref>
          ]) and
natural language queries which can either be a question (e.g., “Are gas prices too high?” from the Touche
2021 Argument Retrieval task [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ]) or a description of an information need (e.g., “Provide information
about the genes Ret and GDNF in kidney development.” from TREC 2005 Genomics Track [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ]). The
5registry.webis.de/code-lib/public-images/query-interpretation:1.0
registry.webis.de/code-lib/public-images/query-entity-linking:1.0
6https://github.com/webis-de/query-entity-linking
7https://github.com/webis-de/query-interpretation
8https://ir-datasets.com/
Natural Language Queries
antique/test [
          <xref ref-type="bibr" rid="ref39">39</xref>
          ] 200
argsme/2020-04-01/touche-2020-task-1 [
          <xref ref-type="bibr" rid="ref40">40, 41</xref>
          ] 49
argsme/2020-04-01/touche-2021-task-1 [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ] 50
clueweb12/touche-2020-task-2 [
          <xref ref-type="bibr" rid="ref40">40, 42, 43</xref>
          ] 50
clueweb12/touche-2021-task-2 [
          <xref ref-type="bibr" rid="ref17">17</xref>
          ] 50
cranfield - 225
medline/2004/trec-genomics-2004 [44] 50
medline/2004/trec-genomics-2005 [
          <xref ref-type="bibr" rid="ref18">18</xref>
          ] 50
msmarco-passage/trec-dl-2019/judged [45, 46] 43
msmarco-passage/trec-dl-2020/judged [47, 46] 54
vaswani - 93
Other
medline/2017/trec-pm-2017
medline/2017/trec-pm-2018
Sum
datasets of the TREC Precision Medicine Track of 2017 [
          <xref ref-type="bibr" rid="ref19">19</xref>
          ] and 2018 [
          <xref ref-type="bibr" rid="ref20">20</xref>
          ] do not fit into either of these
query type categories since the queries from these datasets are compounds of a disease name, a gene,
and a demographic (e.g., “melanoma BRAF (V600E) 64-year-old male”).
        </p>
      </sec>
      <sec id="sec-3-2">
        <title>3.2. Analytics of Entities in Queries</title>
        <p>Table 3 presents query and entity statistics from 31 diferent information retrieval datasets. All datasets
contain between 30 (TREC Precision Medicine Track 2017 dataset) and 325 (NFCorpus test collection)
queries of which almost all queries contain at least one entity according to the query entity linking
approach. Across all datasets, almost 98% of the queries contain at least one entity. The highest ratio of
queries without any entity has the dataset of TREC Web Track 2010. The seven queries from that dataset
for which no entity have been identified are “horse hooves”, “iron”, “vldl levels”, “kiwi”, “tornadoes”,
“raised gardens”, and “ocd”. Except for Obessive-compulsive disorder (OCD), no other likely entity
would be expected in these queries. Although there are songs called “kiwi” or “iron”, neither seems to
be common enough to be included as a plausible entity in an interpretation of these queries.</p>
        <p>A fairly obvious observation is that natural language queries are on average much longer than
keyword queries. While keyword queries consist of approximately three query terms, the average
length of natural language queries vary between 5 and 18 terms depending on the dataset. Intuitively,
the number of mentions of entities correlates with the length of the query. We calculated a Spearman’s
 = 0.63 between these two values over all 2,544 queries, which means that the number of plausible
entities increases monotonically as the length of a query increases. The query with the most entity
mentions is “is it possible to determine rates of forced convective heat transfer from heated cylinders of
non-circular cross-section, (the fluid flow being along the generators)” from the Cranfield dataset. The
query entity linking approach identified 22 diferent segments that might refer to entities in that query.</p>
      </sec>
      <sec id="sec-3-3">
        <title>3.3. Analytics of Query Interpretations</title>
        <p>The query interpretation approach depends on query segmentation, which requires keyword queries and
does not scale well for long natural language queries. We therefore only report on query interpretation
statistics for keyword query datasets. Table 4 presents query interpretation statistics for 18 keyword
query datasets in TIREx. As most statistics are influenced by them, the number of queries and the query
length are copied from Table 3 to facilitate comparisons.</p>
        <p>Overall, the query interpretation approach has identified 2,304 plausible interpretations from 1,225
analyzed queries. Therefore, queries across all datasets have an average of about 1.8 plausible
interpretations using the query interpretation approach, with queries from the TREC Terabyte Track 2006
dataset having the most interpretations of 2.3. One of two queries which have the most interpretations
is “how has african american music influence history” from the TREC Web Track 2014 dataset with
20 diferent interpretations. The highest ranked intepretations with a score close to one are ⟨how |
has | African-American | Music | influence | History⟩ and equivalent variations of that in which
the entities African-American, Music, and History have not been linked, and thus have been
considered concepts. Since the concepts music and history are equivalent to their linked entities, the
interpretation variations can be considered semantically equivalent. Another interesting interpretation
of that query is ⟨how | has | African-American_Music | influence | History⟩ which is probably a
more relevant interpretation because it correctly identifies the relation between “african-american” and
“music”. Unfortunately, this interpretation has been ranked lower.</p>
        <p>Ideally, the more ambiguous a query is (i.e., the more plausible interpretations it has), the more
documents that fulfill all the diferent information needs become relevant. To analyze if this is given in
the datasets, we compute correlation coeficients as Spearman’s  between the number of automatically
identified interpretations and the number of relevant documents (relevance &gt; 0) in the datasets. We have
found no correlation ( ≈ 0) for most datasets or in the worst case a negative correlation ( = − 0.35).
This points into an interesting future direction to use query interpretation as an intermeditate step to
diversify search result to increase this correlation. A related research question would be to analyze
whether each diferent interpretation results in a similar number of relevant documents, which would
lead to a linear growth of relevant documents as the number of interpretations increases.</p>
      </sec>
    </sec>
    <sec id="sec-4">
      <title>4. Conclusions</title>
      <p>To summarize, we contributed query entity linking and query interpretation components to TIREx. A
total of 89,289 Wikipedia-linked entities and 2,304 segmentation-based interpretations were
automatically identified, which can be reproducibly used for future research with the help of TIREx. As part of a
preliminary analysis of the query interpretations, we found that the number of relevant documents for
a query does not correlate with the number of plausible interpretations. This fact points in the direction
of future research in which query interpretation can be used to diversify search results and what other
efects query interpretation has as an intermediate step on an information retrieval pipeline.</p>
    </sec>
    <sec id="sec-5">
      <title>5. Limitations</title>
      <p>Although working with a static entity collection and precomputed commonness scores brings the
advantage of reproducibility, entities that are relevant now may not be as relevant in the future.
Therefore, the identified linked entities included in our query interpretations may become outdated,
or their relevance score may become inadequate. A method to automatically update the entity index
and the associated commonness scores from an up-to-date knowledge base can compensate for this
limitation and will be implemented in the future.</p>
      <p>Wikipedia is a well maintained knowledge base for general knowledge. However, entities from
specialized areas may be inadequately represented or simply not exist on Wikipedia, and consequently
the entity linking method will fail to identify these entities. A mechanism for exchanging (or extending)
the knowledge base that is used for query entity linking can help to find more relevant entities. For
example, the addition of a knowledge base such as PubMed9 could increase the discoverability of
medical-related entities. We aim to implement an easy method to modify and extend the knowledge
base for entity linking and the interpretation of queries.
H. Wachsmuth, M. Potthast, M. Hagen, Overview of touché 2020: Argument retrieval, in: A.
Arampatzis, E. Kanoulas, T. Tsikrika, S. Vrochidis, H. Joho, C. Lioma, C. Eickhof, A. Névéol, L. Cappellato,
N. Ferro (Eds.), Experimental IR Meets Multilinguality, Multimodality, and Interaction. 11th
International Conference of the CLEF Association (CLEF 2020), volume 12260 of Lecture Notes in
Computer Science, Springer, Berlin Heidelberg New York, 2020, pp. 384–395.
[41] H. Wachsmuth, N. Naderi, Y. Hou, Y. Bilu, V. Prabhakaran, T. A. Thijm, G. Hirst, B. Stein,
Computational argumentation quality assessment in natural language, in: P. Blunsom, A. Koller, M. Lapata
(Eds.), 15th Conference of the European Chapter of the Association for Computational Linguistics
(EACL 2017), 2017, pp. 176–187.
[42] L. Braunstain, O. Kurland, D. Carmel, I. Szpektor, A. Shtok, Supporting human answers for
advice-seeking questions in CQA sites, in: N. Ferro, F. Crestani, M.-F. Moens, J. Mothe, F. Silvestri,
G. M. D. Nunzio, C. Hauf, G. Silvello (Eds.), Advances in Information Retrieval - 38th European
Conference on IR Research, ECIR 2016, Padua, Italy, March 20-23, 2016. Proceedings, volume 9626
of Lecture Notes in Computer Science, Springer, 2016, pp. 129–141.
[43] M. Rafalak, K. Abramczuk, A. Wierzbicki, Incredible: Is (almost) all web content trustworthy?
analysis of psychological factors related to website credibility evaluation, in: C.-W. Chung, A. Z.
Broder, K. Shim, T. Suel (Eds.), 23rd International World Wide Web Conference, WWW ’14, Seoul,
Republic of Korea, April 7-11, 2014, Companion Volume, ACM, 2014, pp. 1117–1122.
[44] W. R. Hersh, R. T. Bhuptiraju, L. Ross, P. Johnson, A. M. Cohen, D. F. Kraemer, TREC 2004 genomics
track overview, in: TREC, 2004.
[45] N. Craswell, B. Mitra, E. Yilmaz, D. Campos, E. Voorhees, Overview of the TREC 2019 deep
learning track, in: TREC 2019, 2019.
[46] P. Bajaj, D. Campos, N. Craswell, L. Deng, J. Gao, X. Liu, R. Majumder, A. McNamara, B. Mitra,
T. Nguyen, M. Rosenberg, X. Song, A. Stoica, S. Tiwary, T. Wang, MS MARCO: A Human Generated
MAchine Reading COmprehension Dataset, 2018. arXiv:1611.09268.
[47] N. Craswell, B. Mitra, E. Yilmaz, D. Campos, Overview of the TREC 2020 deep learning track, in:
TREC, 2020.</p>
    </sec>
  </body>
  <back>
    <ref-list>
      <ref id="ref1">
        <mixed-citation>
          [1]
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J. H.</given-names>
            <surname>Reimer</surname>
          </string-name>
          , S. MacAvaney,
          <string-name>
            <given-names>N.</given-names>
            <surname>Deckers</surname>
          </string-name>
          , S. Reich, J.
          <string-name>
            <surname>Bevendorf</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Stein</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Hagen</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Potthast</surname>
          </string-name>
          ,
          <article-title>The Information Retrieval Experiment Platform</article-title>
          , in: H.
          <string-name>
            <surname>-H. Chen</surname>
            ,
            <given-names>W.-J. E.</given-names>
          </string-name>
          <string-name>
            <surname>Duh</surname>
          </string-name>
          , H.
          <string-name>
            <surname>-H. Huang</surname>
            ,
            <given-names>M. P.</given-names>
          </string-name>
          <string-name>
            <surname>Kato</surname>
            ,
            <given-names>J.</given-names>
          </string-name>
          <string-name>
            <surname>Mothe</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          Poblete (Eds.),
          <source>46th International ACM SIGIR Conference on Research and Development in Information Retrieval (SIGIR</source>
          <year>2023</year>
          ), ACM,
          <year>2023</year>
          , pp.
          <fpage>2826</fpage>
          -
          <lpage>2836</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref2">
        <mixed-citation>
          [2]
          <string-name>
            <given-names>S.</given-names>
            <surname>MacAvaney</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Yates</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Feldman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Downey</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Cohan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Goharian</surname>
          </string-name>
          ,
          <article-title>Simplified Data Wrangling with ir_datasets</article-title>
          ,
          <source>in: Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          ,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          , Virtual Event Canada,
          <year>2021</year>
          , pp.
          <fpage>2429</fpage>
          -
          <lpage>2436</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref3">
        <mixed-citation>
          [3]
          <string-name>
            <given-names>S.</given-names>
            <surname>MacAvaney</surname>
          </string-name>
          , C. Macdonald,
          <string-name>
            <surname>I. Ounis</surname>
          </string-name>
          ,
          <article-title>Streamlining Evaluation with ir-measures</article-title>
          , in: M.
          <string-name>
            <surname>Hagen</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Verberne</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Macdonald</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Seifert</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Balog</surname>
            ,
            <given-names>K.</given-names>
          </string-name>
          <string-name>
            <surname>Nørvåg</surname>
          </string-name>
          , V. Setty (Eds.),
          <source>Advances in Information Retrieval</source>
          , volume
          <volume>13186</volume>
          , Springer International Publishing, Cham,
          <year>2022</year>
          , pp.
          <fpage>305</fpage>
          -
          <lpage>310</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref4">
        <mixed-citation>
          [4]
          <string-name>
            <given-names>C.</given-names>
            <surname>Macdonald</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Tonellotto</surname>
          </string-name>
          , S. MacAvaney, I. Ounis,
          <article-title>PyTerrier: Declarative Experimentation in Python from BM25 to Dense Retrieval</article-title>
          ,
          <source>in: Proceedings of the 30th ACM International Conference on Information &amp; Knowledge Management</source>
          ,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          , Virtual Event Queensland Australia,
          <year>2021</year>
          , pp.
          <fpage>4526</fpage>
          -
          <lpage>4533</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref5">
        <mixed-citation>
          [5]
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Wiegmann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Kolyada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Grahm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Elstner</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Loebe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hagen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <article-title>Continuous integration for reproducible shared tasks with TIRA.io</article-title>
          , in: J.
          <string-name>
            <surname>Kamps</surname>
            ,
            <given-names>L.</given-names>
          </string-name>
          <string-name>
            <surname>Goeuriot</surname>
            ,
            <given-names>F.</given-names>
          </string-name>
          <string-name>
            <surname>Crestani</surname>
            ,
            <given-names>M.</given-names>
          </string-name>
          <string-name>
            <surname>Maistro</surname>
            ,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Joho</surname>
            ,
            <given-names>B.</given-names>
          </string-name>
          <string-name>
            <surname>Davis</surname>
            ,
            <given-names>C.</given-names>
          </string-name>
          <string-name>
            <surname>Gurrin</surname>
            ,
            <given-names>U.</given-names>
          </string-name>
          <string-name>
            <surname>Kruschwitz</surname>
            ,
            <given-names>A</given-names>
          </string-name>
          . Caputo (Eds.),
          <source>Advances in Information Retrieval. 45th European Conference on IR Research (ECIR</source>
          <year>2023</year>
          ), Lecture Notes in Computer Science, Springer, Berlin Heidelberg New York,
          <year>2023</year>
          , pp.
          <fpage>236</fpage>
          -
          <lpage>241</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref6">
        <mixed-citation>
          [6]
          <string-name>
            <given-names>V.</given-names>
            <surname>Kasturia</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Gohsen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hagen</surname>
          </string-name>
          ,
          <article-title>Query Interpretations from Entity-Linked Segmentations</article-title>
          ,
          <source>in: 15th ACM International Conference on Web Search and Data Mining (WSDM</source>
          <year>2022</year>
          ), ACM,
          <year>2022</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref7">
        <mixed-citation>
          [7]
          <string-name>
            <given-names>D.</given-names>
            <surname>Shehata</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Arabzadeh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. L. A.</given-names>
            <surname>Clarke</surname>
          </string-name>
          ,
          <article-title>Early Stage Sparse Retrieval with Entity Linking</article-title>
          ,
          <source>in: Proceedings of the 31st ACM International Conference on Information &amp; Knowledge Management</source>
          ,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          ,
          <source>Atlanta GA USA</source>
          ,
          <year>2022</year>
          , pp.
          <fpage>4464</fpage>
          -
          <lpage>4469</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref8">
        <mixed-citation>
          [8]
          <string-name>
            <given-names>F.</given-names>
            <surname>Hasibi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Balog</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S. E.</given-names>
            <surname>Bratsberg</surname>
          </string-name>
          ,
          <article-title>Exploiting Entity Linking in Queries for Entity Retrieval</article-title>
          ,
          <source>in: Proceedings of the 2016 ACM International Conference on the Theory of Information Retrieval</source>
          ,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          , Newark Delaware USA,
          <year>2016</year>
          , pp.
          <fpage>209</fpage>
          -
          <lpage>218</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref9">
        <mixed-citation>
          [9]
          <string-name>
            <given-names>S.</given-names>
            <surname>Chatterjee</surname>
          </string-name>
          , L. Dietz,
          <article-title>Entity Retrieval Using Fine-Grained Entity Aspects</article-title>
          ,
          <source>in: Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval</source>
          ,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          , Virtual Event Canada,
          <year>2021</year>
          , pp.
          <fpage>1662</fpage>
          -
          <lpage>1666</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref10">
        <mixed-citation>
          [10]
          <string-name>
            <surname>M. L. Sidi</surname>
            ,
            <given-names>S.</given-names>
          </string-name>
          <string-name>
            <surname>Gunal</surname>
            ,
            <given-names>A Purely</given-names>
          </string-name>
          <string-name>
            <surname>Entity-Based Semantic</surname>
          </string-name>
          Search Approach for Document Retrieval,
          <source>Applied Sciences</source>
          <volume>13</volume>
          (
          <year>2023</year>
          )
          <fpage>10285</fpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref11">
        <mixed-citation>
          [11]
          <string-name>
            <given-names>P.</given-names>
            <surname>Ferragina</surname>
          </string-name>
          , U. Scaiella, TAGME:
          <article-title>On-the-fly annotation of short text fragments (by wikipedia entities)</article-title>
          ,
          <source>in: Proceedings of the 19th ACM International Conference on Information and Knowledge Management</source>
          ,
          <string-name>
            <surname>ACM</surname>
          </string-name>
          , Toronto ON Canada,
          <year>2010</year>
          , pp.
          <fpage>1625</fpage>
          -
          <lpage>1628</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref12">
        <mixed-citation>
          [12]
          <string-name>
            <given-names>C.</given-names>
            <surname>Clarke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Craswell</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Soborof</surname>
          </string-name>
          ,
          <article-title>Overview of the TREC 2009 web track</article-title>
          , in: E. M.
          <string-name>
            <surname>Voorhees</surname>
            ,
            <given-names>L. P.</given-names>
          </string-name>
          Buckland (Eds.),
          <source>Proceedings of the Eighteenth Text REtrieval Conference</source>
          , TREC 2009, Gaithersburg, Maryland, USA, November
          <volume>17</volume>
          -
          <issue>20</issue>
          ,
          <year>2009</year>
          , volume
          <volume>500</volume>
          -278 of NIST Special Publication,
          <source>National Institute of Standards and Technology (NIST)</source>
          ,
          <year>2009</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref13">
        <mixed-citation>
          [13]
          <string-name>
            <given-names>C.</given-names>
            <surname>Clarke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Craswell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Voorhees</surname>
          </string-name>
          ,
          <article-title>Overview of the TREC 2012 web track</article-title>
          , in: E. M.
          <string-name>
            <surname>Voorhees</surname>
            ,
            <given-names>L. P.</given-names>
          </string-name>
          Buckland (Eds.),
          <source>Proceedings of the Twenty-First Text REtrieval Conference</source>
          , TREC 2012, Gaithersburg, Maryland, USA, November 6-
          <issue>9</issue>
          ,
          <year>2012</year>
          , volume
          <volume>500</volume>
          -298 of NIST Special Publication,
          <source>National Institute of Standards and Technology (NIST)</source>
          ,
          <year>2012</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref14">
        <mixed-citation>
          [14]
          <string-name>
            <given-names>M.</given-names>
            <surname>Hagen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Bräutigam</surname>
          </string-name>
          ,
          <article-title>Query segmentation revisited</article-title>
          , in: S. Srinivasan,
          <string-name>
            <given-names>K.</given-names>
            <surname>Ramamritham</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Kumar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M. P.</given-names>
            <surname>Ravindra</surname>
          </string-name>
          , E. Bertino, R. Kumar (Eds.),
          <source>20th International Conference on World Wide Web (WWW</source>
          <year>2011</year>
          ), ACM,
          <year>2011</year>
          , pp.
          <fpage>97</fpage>
          -
          <lpage>106</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref15">
        <mixed-citation>
          [15]
          <string-name>
            <given-names>M.</given-names>
            <surname>Hagen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Beyer</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <article-title>Towards optimum query segmentation: In doubt without</article-title>
          , in: X.
          <string-name>
            <surname>Chen</surname>
            , G. Lebanon,
            <given-names>H.</given-names>
          </string-name>
          <string-name>
            <surname>Wang</surname>
            ,
            <given-names>M. J.</given-names>
          </string-name>
          <string-name>
            <surname>Zaki</surname>
          </string-name>
          (Eds.),
          <source>21st ACM International Conference on Information and Knowledge Management (CIKM</source>
          <year>2012</year>
          ), ACM,
          <year>2012</year>
          , pp.
          <fpage>1015</fpage>
          -
          <lpage>1024</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref16">
        <mixed-citation>
          [16]
          <string-name>
            <given-names>I.</given-names>
            <surname>Yamada</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Shindo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Takeda</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Takefuji</surname>
          </string-name>
          ,
          <article-title>Joint Learning of the Embedding of Words and Entities for Named Entity Disambiguation</article-title>
          ,
          <source>in: Proceedings of The 20th SIGNLL Conference on Computational Natural Language Learning, Association for Computational Linguistics</source>
          , Berlin, Germany,
          <year>2016</year>
          , pp.
          <fpage>250</fpage>
          -
          <lpage>259</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref17">
        <mixed-citation>
          [17]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bondarenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Gienapp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Beloucif</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ajjour</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Panchenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Biemann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Wachsmuth</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Potthast</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hagen</surname>
          </string-name>
          , Overview of touché 2021:
          <article-title>Argument retrieval</article-title>
          , in: K. S. Candan,
          <string-name>
            <given-names>B.</given-names>
            <surname>Ionescu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Goeuriot</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Müller</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Joly</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Maistro</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Piroi</surname>
          </string-name>
          , G. Faggioli, N. Ferro (Eds.),
          <source>Experimental IR Meets Multilinguality, Multimodality, and Interaction. 12th International Conference of the CLEF Association (CLEF</source>
          <year>2021</year>
          ), volume
          <volume>12880</volume>
          of Lecture Notes in Computer Science, Springer, Berlin Heidelberg New York,
          <year>2021</year>
          , pp.
          <fpage>450</fpage>
          -
          <lpage>467</lpage>
          .
        </mixed-citation>
      </ref>
      <ref id="ref18">
        <mixed-citation>
          [18]
          <string-name>
            <given-names>W.</given-names>
            <surname>Hersh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Cohen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Roberts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Hearst</surname>
          </string-name>
          ,
          <article-title>TREC 2005 Genomics Track Overview</article-title>
          , in
          <source>: Proceedings of the Fourteenth Text REtrieval Conference (TREC</source>
          <year>2005</year>
          ),
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref19">
        <mixed-citation>
          [19]
          <string-name>
            <given-names>K.</given-names>
            <surname>Roberts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Demner-Fushman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Voorhees</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. R.</given-names>
            <surname>Hersh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bedrick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. J.</given-names>
            <surname>Lazar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Pant</surname>
          </string-name>
          ,
          <article-title>Overview of the TREC 2017 Precision Medicine Track</article-title>
          , TREC
          <volume>26</volume>
          (
          <year>2017</year>
          ) https://trec.nist.gov/pubs/trec26/papers/Overview-PM.pdf.
        </mixed-citation>
      </ref>
      <ref id="ref20">
        <mixed-citation>
          [20]
          <string-name>
            <given-names>K.</given-names>
            <surname>Roberts</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Demner-Fushman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E.</given-names>
            <surname>Vorhees</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W. R.</given-names>
            <surname>Hersh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bedrick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A. J.</given-names>
            <surname>Lazar</surname>
          </string-name>
          ,
          <article-title>Overview of the TREC 2018 Precision Medicine Track</article-title>
          ,
          <string-name>
            <surname>TREC</surname>
          </string-name>
          (
          <year>2018</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref21">
        <mixed-citation>
          [21]
          <string-name>
            <given-names>C.</given-names>
            <surname>Clarke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Craswell</surname>
          </string-name>
          , I. Soborof,
          <string-name>
            <given-names>G. V.</given-names>
            <surname>Cormack</surname>
          </string-name>
          ,
          <article-title>Overview of the TREC 2010 web track</article-title>
          ,
          <source>in: TREC</source>
          ,
          <year>2010</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref22">
        <mixed-citation>
          [22]
          <string-name>
            <given-names>C.</given-names>
            <surname>Clarke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Craswell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>I.</given-names>
            <surname>Soborof</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Voorhees</surname>
          </string-name>
          ,
          <article-title>Overview of the TREC 2011 web track</article-title>
          ,
          <source>in: TREC</source>
          ,
          <year>2011</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref23">
        <mixed-citation>
          [23]
          <string-name>
            <given-names>K.</given-names>
            <surname>Collins-Thompson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bennett</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Diaz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. L. A.</given-names>
            <surname>Clarke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Voorhees</surname>
          </string-name>
          ,
          <article-title>TREC 2013 web track overview</article-title>
          , in: TREC,
          <year>2013</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref24">
        <mixed-citation>
          [24]
          <string-name>
            <given-names>K.</given-names>
            <surname>Collins-Thompson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Macdonald</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Bennett</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Diaz</surname>
          </string-name>
          ,
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Voorhees</surname>
          </string-name>
          ,
          <article-title>TREC 2014 web track overview</article-title>
          , in: TREC,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref25">
        <mixed-citation>
          [25]
          <string-name>
            <given-names>E.</given-names>
            <surname>Voorhees</surname>
          </string-name>
          ,
          <string-name>
            <given-names>T.</given-names>
            <surname>Alam</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Bedrick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Demner-Fushman</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Hersh</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Roberts</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Soborof</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L. L.</given-names>
            <surname>Wang</surname>
          </string-name>
          , TREC-COVID:
          <article-title>Constructing a pandemic information retrieval test collection</article-title>
          , ArXiv abs/
          <year>2005</year>
          .04474 (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref26">
        <mixed-citation>
          [26]
          <string-name>
            <given-names>L. L.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Lo</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Chandrasekhar</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Reas</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Yang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Eide</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Funk</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R. M.</given-names>
            <surname>Kinney</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Liu</surname>
          </string-name>
          ,
          <string-name>
            <given-names>W.</given-names>
            <surname>Merrill</surname>
          </string-name>
          ,
          <string-name>
            <given-names>P.</given-names>
            <surname>Mooney</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Murdick</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Rishi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>J.</given-names>
            <surname>Sheehan</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Z.</given-names>
            <surname>Shen</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stilson</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Wade</surname>
          </string-name>
          ,
          <string-name>
            <given-names>K.</given-names>
            <surname>Wang</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Wilhelm</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Xie</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Raymond</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D. S.</given-names>
            <surname>Weld</surname>
          </string-name>
          ,
          <string-name>
            <given-names>O.</given-names>
            <surname>Etzioni</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Kohlmeier</surname>
          </string-name>
          , CORD-
          <volume>19</volume>
          : The covid-19 open research dataset,
          <source>ArXiv</source>
          (
          <year>2020</year>
          ).
        </mixed-citation>
      </ref>
      <ref id="ref27">
        <mixed-citation>
          [27]
          <string-name>
            <given-names>S.</given-names>
            <surname>Huston</surname>
          </string-name>
          , W. B.
          <string-name>
            <surname>Croft</surname>
          </string-name>
          ,
          <article-title>A comparison of retrieval models using term dependencies</article-title>
          ,
          <source>in: CIKM</source>
          ,
          <year>2014</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref28">
        <mixed-citation>
          [28]
          <string-name>
            <given-names>E.</given-names>
            <surname>Voorhees</surname>
          </string-name>
          ,
          <article-title>Overview of the TREC 2004 robust retrieval track</article-title>
          ,
          <source>in: TREC</source>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref29">
        <mixed-citation>
          [29]
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Voorhees</surname>
          </string-name>
          ,
          <string-name>
            <surname>NIST</surname>
          </string-name>
          <article-title>TREC disks 4 and 5: Retrieval test collections document set</article-title>
          ,
          <year>1996</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref30">
        <mixed-citation>
          [30]
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Voorhees</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Harman</surname>
          </string-name>
          ,
          <article-title>Overview of the seventh text retrieval conference (TREC-7)</article-title>
          , in: TREC,
          <year>1998</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref31">
        <mixed-citation>
          [31]
          <string-name>
            <given-names>E. M.</given-names>
            <surname>Voorhees</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Harman</surname>
          </string-name>
          ,
          <article-title>Overview of the eight text retrieval conference (TREC-8)</article-title>
          , in: TREC,
          <year>1999</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref32">
        <mixed-citation>
          [32]
          <string-name>
            <given-names>N.</given-names>
            <surname>Craswell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Hawking</surname>
          </string-name>
          ,
          <article-title>Overview of the TREC-2002 web track</article-title>
          ,
          <source>in: TREC</source>
          ,
          <year>2002</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref33">
        <mixed-citation>
          [33]
          <string-name>
            <given-names>N.</given-names>
            <surname>Craswell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Hawking</surname>
          </string-name>
          ,
          <string-name>
            <given-names>R.</given-names>
            <surname>Wilkinson</surname>
          </string-name>
          ,
          <string-name>
            <surname>M.</surname>
          </string-name>
          <article-title>Wu, Overview of the TREC 2003 web track</article-title>
          ,
          <source>in: TREC</source>
          ,
          <year>2003</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref34">
        <mixed-citation>
          [34]
          <string-name>
            <given-names>N.</given-names>
            <surname>Craswell</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Hawking</surname>
          </string-name>
          ,
          <article-title>Overview of the TREC-2004 web track</article-title>
          ,
          <source>in: TREC</source>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref35">
        <mixed-citation>
          [35]
          <string-name>
            <given-names>C.</given-names>
            <surname>Clarke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>N.</given-names>
            <surname>Craswell</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Soborof</surname>
          </string-name>
          ,
          <article-title>Overview of the TREC 2004 terabyte track</article-title>
          ,
          <source>in: TREC</source>
          ,
          <year>2004</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref36">
        <mixed-citation>
          [36]
          <string-name>
            <given-names>C. L. A.</given-names>
            <surname>Clarke</surname>
          </string-name>
          ,
          <string-name>
            <given-names>F.</given-names>
            <surname>Scholer</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Soborof</surname>
          </string-name>
          ,
          <article-title>The TREC 2005 terabyte track</article-title>
          ,
          <source>in: TREC</source>
          ,
          <year>2005</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref37">
        <mixed-citation>
          [37]
          <string-name>
            <given-names>S.</given-names>
            <surname>Büttcher</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C. L. A.</given-names>
            <surname>Clarke</surname>
          </string-name>
          ,
          <string-name>
            <surname>I. Soborof</surname>
          </string-name>
          ,
          <article-title>The TREC 2006 terabyte track</article-title>
          ,
          <source>in: TREC</source>
          ,
          <year>2006</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref38">
        <mixed-citation>
          [38]
          <string-name>
            <given-names>V.</given-names>
            <surname>Boteva</surname>
          </string-name>
          ,
          <string-name>
            <given-names>D.</given-names>
            <surname>Gholipour</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Sokolov</surname>
          </string-name>
          ,
          <string-name>
            <given-names>S.</given-names>
            <surname>Riezler</surname>
          </string-name>
          ,
          <article-title>A full-text learning to rank dataset for medical information retrieval</article-title>
          ,
          <source>in: Proceedings of the European Conference on Information Retrieval (ECIR)</source>
          , Springer, Padova, Italy,
          <year>2016</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref39">
        <mixed-citation>
          [39]
          <string-name>
            <given-names>H.</given-names>
            <surname>Hashemi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Aliannejadi</surname>
          </string-name>
          ,
          <string-name>
            <given-names>H.</given-names>
            <surname>Zamani</surname>
          </string-name>
          ,
          <string-name>
            <surname>B. Croft,</surname>
          </string-name>
          <article-title>ANTIQUE: A non-factoid question answering benchmark</article-title>
          ,
          <source>in: ECIR</source>
          ,
          <year>2020</year>
          .
        </mixed-citation>
      </ref>
      <ref id="ref40">
        <mixed-citation>
          [40]
          <string-name>
            <given-names>A.</given-names>
            <surname>Bondarenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Fröbe</surname>
          </string-name>
          ,
          <string-name>
            <given-names>M.</given-names>
            <surname>Beloucif</surname>
          </string-name>
          ,
          <string-name>
            <given-names>L.</given-names>
            <surname>Gienapp</surname>
          </string-name>
          ,
          <string-name>
            <given-names>Y.</given-names>
            <surname>Ajjour</surname>
          </string-name>
          ,
          <string-name>
            <given-names>A.</given-names>
            <surname>Panchenko</surname>
          </string-name>
          ,
          <string-name>
            <given-names>C.</given-names>
            <surname>Biemann</surname>
          </string-name>
          ,
          <string-name>
            <given-names>B.</given-names>
            <surname>Stein</surname>
          </string-name>
          ,
        </mixed-citation>
      </ref>
    </ref-list>
  </back>
</article>